diff --git a/.gitattributes b/.gitattributes index 6844c60fa39661f7ab631812caa048af54fbb2d9..8b8986bd98e55db5da6185c93db8f201bfc95a86 100644 --- a/.gitattributes +++ b/.gitattributes @@ -110,3 +110,7 @@ venv/lib/python3.10/site-packages/torch/lib/libtorch_cpu.so filter=lfs diff=lfs venv/bin/python filter=lfs diff=lfs merge=lfs -text venv/lib/python3.10/site-packages/nvidia/cusparse/lib/libcusparse.so.12 filter=lfs diff=lfs merge=lfs -text venv/lib/python3.10/site-packages/nvidia/nccl/lib/libnccl.so.2 filter=lfs diff=lfs merge=lfs -text +lm-evaluation-harness/wandb/run-20240608_134753-jc54oi3l/run-jc54oi3l.wandb filter=lfs diff=lfs merge=lfs -text +lm-evaluation-harness/wandb/run-20240605_075207-j37ty43h/run-j37ty43h.wandb filter=lfs diff=lfs merge=lfs -text +lm-evaluation-harness/wandb/run-20240608_111026-9apxn9eo/run-9apxn9eo.wandb filter=lfs diff=lfs merge=lfs -text +lm-evaluation-harness/wandb/run-20240608_190333-82mnef5m/run-82mnef5m.wandb filter=lfs diff=lfs merge=lfs -text diff --git a/lm-evaluation-harness/wandb/run-20240514_103002-s6mpkihm/logs/debug-internal.log b/lm-evaluation-harness/wandb/run-20240514_103002-s6mpkihm/logs/debug-internal.log new file mode 100644 index 0000000000000000000000000000000000000000..b60ed955060abccf132fe83b5a0b9d817569ddd4 --- /dev/null +++ b/lm-evaluation-harness/wandb/run-20240514_103002-s6mpkihm/logs/debug-internal.log @@ -0,0 +1,190 @@ +2024-05-14 10:30:02,053 INFO StreamThr :1843 [internal.py:wandb_internal():85] W&B internal server running at pid: 1843, started at: 2024-05-14 10:30:02.052986 +2024-05-14 10:30:02,055 DEBUG HandlerThread:1843 [handler.py:handle_request():158] handle_request: status +2024-05-14 10:30:02,056 INFO WriterThread:1843 [datastore.py:open_for_write():87] open: /data/cronscript/lm-evaluation-harness/wandb/run-20240514_103002-s6mpkihm/run-s6mpkihm.wandb +2024-05-14 10:30:02,057 DEBUG SenderThread:1843 [sender.py:send():378] send: header +2024-05-14 10:30:02,066 DEBUG SenderThread:1843 [sender.py:send():378] send: run +2024-05-14 10:30:02,296 INFO SenderThread:1843 [dir_watcher.py:__init__():211] watching files in: /data/cronscript/lm-evaluation-harness/wandb/run-20240514_103002-s6mpkihm/files +2024-05-14 10:30:02,296 INFO SenderThread:1843 [sender.py:_start_run_threads():1123] run started: s6mpkihm with start time 1715682602.052393 +2024-05-14 10:30:02,302 DEBUG HandlerThread:1843 [handler.py:handle_request():158] handle_request: check_version +2024-05-14 10:30:02,302 DEBUG SenderThread:1843 [sender.py:send_request():405] send_request: check_version +2024-05-14 10:30:02,382 DEBUG HandlerThread:1843 [handler.py:handle_request():158] handle_request: run_start +2024-05-14 10:30:02,383 DEBUG HandlerThread:1843 [system_info.py:__init__():26] System info init +2024-05-14 10:30:02,383 DEBUG HandlerThread:1843 [system_info.py:__init__():41] System info init done +2024-05-14 10:30:02,383 INFO HandlerThread:1843 [system_monitor.py:start():194] Starting system monitor +2024-05-14 10:30:02,383 INFO SystemMonitor:1843 [system_monitor.py:_start():158] Starting system asset monitoring threads +2024-05-14 10:30:02,384 INFO HandlerThread:1843 [system_monitor.py:probe():214] Collecting system info +2024-05-14 10:30:02,384 INFO SystemMonitor:1843 [interfaces.py:start():188] Started cpu monitoring +2024-05-14 10:30:02,384 INFO SystemMonitor:1843 [interfaces.py:start():188] Started disk monitoring +2024-05-14 10:30:02,384 INFO SystemMonitor:1843 [interfaces.py:start():188] Started memory monitoring +2024-05-14 10:30:02,385 INFO SystemMonitor:1843 [interfaces.py:start():188] Started network monitoring +2024-05-14 10:30:02,466 DEBUG HandlerThread:1843 [system_info.py:probe():150] Probing system +2024-05-14 10:30:02,474 DEBUG HandlerThread:1843 [system_info.py:_probe_git():135] Probing git +2024-05-14 10:30:02,494 ERROR HandlerThread:1843 [gitlib.py:root():92] git root error: Cmd('git') failed due to: exit code(128) + cmdline: git rev-parse --show-toplevel + stderr: 'fatal: detected dubious ownership in repository at '/data/cronscript/lm-evaluation-harness' +To add an exception for this directory, call: + + git config --global --add safe.directory /data/cronscript/lm-evaluation-harness' +2024-05-14 10:30:02,494 DEBUG HandlerThread:1843 [system_info.py:_probe_git():143] Probing git done +2024-05-14 10:30:02,494 DEBUG HandlerThread:1843 [system_info.py:probe():198] Probing system done +2024-05-14 10:30:02,494 DEBUG HandlerThread:1843 [system_monitor.py:probe():223] {'os': 'Linux-5.15.0-92-generic-x86_64-with-glibc2.35', 'python': '3.10.12', 'heartbeatAt': '2024-05-14T10:30:02.466481', 'startedAt': '2024-05-14T10:30:02.042367', 'docker': None, 'cuda': None, 'args': ('--model', 'hf', '--model_args', 'pretrained=/data/cronscript/ckpts//hf_ckpt//global_step100', '--tasks', 'indiccopa-hi', '--batch_size', 'auto', '--wandb_args', 'project=bharatgpt'), 'state': 'running', 'program': '-m lm_eval.__main__', 'codePathLocal': None, 'git': {'remote': 'https://github.com/EleutherAI/lm-evaluation-harness', 'commit': None}, 'email': None, 'root': '/data/cronscript/lm-evaluation-harness', 'host': 'vizzhy-150-3', 'username': 'root', 'executable': '/usr/bin/python3', 'cpu_count': 76, 'cpu_count_logical': 152, 'cpu_freq': {'current': 3392.375447368421, 'min': 800.0, 'max': 3400.0}, 'cpu_freq_per_core': [{'current': 3300.0, 'min': 800.0, 'max': 3400.0}, {'current': 3300.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3300.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3300.003, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3300.0, 'min': 800.0, 'max': 3400.0}, {'current': 3300.0, 'min': 800.0, 'max': 3400.0}, {'current': 3300.003, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3300.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3300.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}], 'disk': {'/': {'total': 866.4415092468262, 'used': 76.92179107666016}}, 'memory': {'total': 1007.5000267028809}} +2024-05-14 10:30:02,494 INFO HandlerThread:1843 [system_monitor.py:probe():224] Finished collecting system info +2024-05-14 10:30:02,494 INFO HandlerThread:1843 [system_monitor.py:probe():227] Publishing system info +2024-05-14 10:30:02,495 INFO HandlerThread:1843 [system_monitor.py:probe():229] Finished publishing system info +2024-05-14 10:30:02,500 DEBUG SenderThread:1843 [sender.py:send():378] send: files +2024-05-14 10:30:02,500 INFO SenderThread:1843 [sender.py:_save_file():1389] saving file wandb-metadata.json with policy now +2024-05-14 10:30:02,596 DEBUG HandlerThread:1843 [handler.py:handle_request():158] handle_request: python_packages +2024-05-14 10:30:02,597 DEBUG SenderThread:1843 [sender.py:send_request():405] send_request: python_packages +2024-05-14 10:30:02,597 DEBUG HandlerThread:1843 [handler.py:handle_request():158] handle_request: stop_status +2024-05-14 10:30:02,598 DEBUG SenderThread:1843 [sender.py:send_request():405] send_request: stop_status +2024-05-14 10:30:02,724 DEBUG SenderThread:1843 [sender.py:send():378] send: telemetry +2024-05-14 10:30:03,015 INFO wandb-upload_0:1843 [upload_job.py:push():130] Uploaded file /tmp/tmpkvz23fo_wandb/nr8ewnt7-wandb-metadata.json +2024-05-14 10:30:03,297 INFO Thread-12 :1843 [dir_watcher.py:_on_file_created():271] file/dir created: /data/cronscript/lm-evaluation-harness/wandb/run-20240514_103002-s6mpkihm/files/wandb-metadata.json +2024-05-14 10:30:03,297 INFO Thread-12 :1843 [dir_watcher.py:_on_file_created():271] file/dir created: /data/cronscript/lm-evaluation-harness/wandb/run-20240514_103002-s6mpkihm/files/output.log +2024-05-14 10:30:03,298 INFO Thread-12 :1843 [dir_watcher.py:_on_file_created():271] file/dir created: /data/cronscript/lm-evaluation-harness/wandb/run-20240514_103002-s6mpkihm/files/requirements.txt +2024-05-14 10:30:05,298 INFO Thread-12 :1843 [dir_watcher.py:_on_file_modified():288] file/dir modified: /data/cronscript/lm-evaluation-harness/wandb/run-20240514_103002-s6mpkihm/files/output.log +2024-05-14 10:30:07,081 DEBUG HandlerThread:1843 [handler.py:handle_request():158] handle_request: status_report +2024-05-14 10:30:09,305 INFO Thread-12 :1843 [dir_watcher.py:_on_file_modified():288] file/dir modified: /data/cronscript/lm-evaluation-harness/wandb/run-20240514_103002-s6mpkihm/files/output.log +2024-05-14 10:30:12,082 DEBUG HandlerThread:1843 [handler.py:handle_request():158] handle_request: status_report +2024-05-14 10:30:15,313 INFO Thread-12 :1843 [dir_watcher.py:_on_file_modified():288] file/dir modified: /data/cronscript/lm-evaluation-harness/wandb/run-20240514_103002-s6mpkihm/files/output.log +2024-05-14 10:30:17,234 DEBUG HandlerThread:1843 [handler.py:handle_request():158] handle_request: status_report +2024-05-14 10:30:17,314 INFO Thread-12 :1843 [dir_watcher.py:_on_file_modified():288] file/dir modified: /data/cronscript/lm-evaluation-harness/wandb/run-20240514_103002-s6mpkihm/files/output.log +2024-05-14 10:30:17,597 DEBUG HandlerThread:1843 [handler.py:handle_request():158] handle_request: stop_status +2024-05-14 10:30:17,597 DEBUG SenderThread:1843 [sender.py:send_request():405] send_request: stop_status +2024-05-14 10:30:19,316 INFO Thread-12 :1843 [dir_watcher.py:_on_file_modified():288] file/dir modified: /data/cronscript/lm-evaluation-harness/wandb/run-20240514_103002-s6mpkihm/files/output.log +2024-05-14 10:30:22,684 DEBUG HandlerThread:1843 [handler.py:handle_request():158] handle_request: status_report +2024-05-14 10:30:23,941 DEBUG SenderThread:1843 [sender.py:send():378] send: exit +2024-05-14 10:30:23,942 INFO SenderThread:1843 [sender.py:send_exit():585] handling exit code: 0 +2024-05-14 10:30:23,942 INFO SenderThread:1843 [sender.py:send_exit():587] handling runtime: 21 +2024-05-14 10:30:23,943 INFO SenderThread:1843 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end +2024-05-14 10:30:23,943 INFO SenderThread:1843 [sender.py:send_exit():593] send defer +2024-05-14 10:30:23,943 DEBUG HandlerThread:1843 [handler.py:handle_request():158] handle_request: defer +2024-05-14 10:30:23,943 INFO HandlerThread:1843 [handler.py:handle_request_defer():184] handle defer: 0 +2024-05-14 10:30:23,943 DEBUG SenderThread:1843 [sender.py:send_request():405] send_request: defer +2024-05-14 10:30:23,943 INFO SenderThread:1843 [sender.py:send_request_defer():609] handle sender defer: 0 +2024-05-14 10:30:23,943 INFO SenderThread:1843 [sender.py:transition_state():613] send defer: 1 +2024-05-14 10:30:23,944 DEBUG HandlerThread:1843 [handler.py:handle_request():158] handle_request: defer +2024-05-14 10:30:23,944 INFO HandlerThread:1843 [handler.py:handle_request_defer():184] handle defer: 1 +2024-05-14 10:30:23,944 DEBUG SenderThread:1843 [sender.py:send_request():405] send_request: defer +2024-05-14 10:30:23,944 INFO SenderThread:1843 [sender.py:send_request_defer():609] handle sender defer: 1 +2024-05-14 10:30:23,944 INFO SenderThread:1843 [sender.py:transition_state():613] send defer: 2 +2024-05-14 10:30:23,944 DEBUG HandlerThread:1843 [handler.py:handle_request():158] handle_request: defer +2024-05-14 10:30:23,944 INFO HandlerThread:1843 [handler.py:handle_request_defer():184] handle defer: 2 +2024-05-14 10:30:23,944 INFO HandlerThread:1843 [system_monitor.py:finish():203] Stopping system monitor +2024-05-14 10:30:23,944 DEBUG SystemMonitor:1843 [system_monitor.py:_start():172] Starting system metrics aggregation loop +2024-05-14 10:30:23,945 INFO HandlerThread:1843 [interfaces.py:finish():200] Joined cpu monitor +2024-05-14 10:30:23,945 DEBUG SystemMonitor:1843 [system_monitor.py:_start():179] Finished system metrics aggregation loop +2024-05-14 10:30:23,945 INFO HandlerThread:1843 [interfaces.py:finish():200] Joined disk monitor +2024-05-14 10:30:23,945 DEBUG SystemMonitor:1843 [system_monitor.py:_start():183] Publishing last batch of metrics +2024-05-14 10:30:23,945 INFO HandlerThread:1843 [interfaces.py:finish():200] Joined memory monitor +2024-05-14 10:30:23,947 INFO HandlerThread:1843 [interfaces.py:finish():200] Joined network monitor +2024-05-14 10:30:23,948 DEBUG SenderThread:1843 [sender.py:send_request():405] send_request: defer +2024-05-14 10:30:23,948 INFO SenderThread:1843 [sender.py:send_request_defer():609] handle sender defer: 2 +2024-05-14 10:30:23,948 INFO SenderThread:1843 [sender.py:transition_state():613] send defer: 3 +2024-05-14 10:30:23,948 DEBUG SenderThread:1843 [sender.py:send():378] send: stats +2024-05-14 10:30:23,949 DEBUG HandlerThread:1843 [handler.py:handle_request():158] handle_request: defer +2024-05-14 10:30:23,949 INFO HandlerThread:1843 [handler.py:handle_request_defer():184] handle defer: 3 +2024-05-14 10:30:23,949 DEBUG SenderThread:1843 [sender.py:send_request():405] send_request: defer +2024-05-14 10:30:23,949 INFO SenderThread:1843 [sender.py:send_request_defer():609] handle sender defer: 3 +2024-05-14 10:30:23,949 INFO SenderThread:1843 [sender.py:transition_state():613] send defer: 4 +2024-05-14 10:30:23,949 DEBUG HandlerThread:1843 [handler.py:handle_request():158] handle_request: defer +2024-05-14 10:30:23,949 INFO HandlerThread:1843 [handler.py:handle_request_defer():184] handle defer: 4 +2024-05-14 10:30:23,949 DEBUG SenderThread:1843 [sender.py:send_request():405] send_request: defer +2024-05-14 10:30:23,949 INFO SenderThread:1843 [sender.py:send_request_defer():609] handle sender defer: 4 +2024-05-14 10:30:23,949 INFO SenderThread:1843 [sender.py:transition_state():613] send defer: 5 +2024-05-14 10:30:23,949 DEBUG HandlerThread:1843 [handler.py:handle_request():158] handle_request: defer +2024-05-14 10:30:23,949 INFO HandlerThread:1843 [handler.py:handle_request_defer():184] handle defer: 5 +2024-05-14 10:30:23,950 DEBUG SenderThread:1843 [sender.py:send():378] send: summary +2024-05-14 10:30:23,950 INFO SenderThread:1843 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end +2024-05-14 10:30:23,950 DEBUG SenderThread:1843 [sender.py:send_request():405] send_request: defer +2024-05-14 10:30:23,950 INFO SenderThread:1843 [sender.py:send_request_defer():609] handle sender defer: 5 +2024-05-14 10:30:23,950 INFO SenderThread:1843 [sender.py:transition_state():613] send defer: 6 +2024-05-14 10:30:23,951 DEBUG HandlerThread:1843 [handler.py:handle_request():158] handle_request: defer +2024-05-14 10:30:23,951 INFO HandlerThread:1843 [handler.py:handle_request_defer():184] handle defer: 6 +2024-05-14 10:30:23,951 DEBUG SenderThread:1843 [sender.py:send_request():405] send_request: defer +2024-05-14 10:30:23,951 INFO SenderThread:1843 [sender.py:send_request_defer():609] handle sender defer: 6 +2024-05-14 10:30:23,953 DEBUG HandlerThread:1843 [handler.py:handle_request():158] handle_request: status_report +2024-05-14 10:30:24,086 INFO SenderThread:1843 [sender.py:transition_state():613] send defer: 7 +2024-05-14 10:30:24,086 DEBUG HandlerThread:1843 [handler.py:handle_request():158] handle_request: defer +2024-05-14 10:30:24,086 INFO HandlerThread:1843 [handler.py:handle_request_defer():184] handle defer: 7 +2024-05-14 10:30:24,087 DEBUG SenderThread:1843 [sender.py:send_request():405] send_request: defer +2024-05-14 10:30:24,087 INFO SenderThread:1843 [sender.py:send_request_defer():609] handle sender defer: 7 +2024-05-14 10:30:24,320 INFO Thread-12 :1843 [dir_watcher.py:_on_file_modified():288] file/dir modified: /data/cronscript/lm-evaluation-harness/wandb/run-20240514_103002-s6mpkihm/files/config.yaml +2024-05-14 10:30:24,320 INFO Thread-12 :1843 [dir_watcher.py:_on_file_created():271] file/dir created: /data/cronscript/lm-evaluation-harness/wandb/run-20240514_103002-s6mpkihm/files/wandb-summary.json +2024-05-14 10:30:24,941 DEBUG HandlerThread:1843 [handler.py:handle_request():158] handle_request: poll_exit +2024-05-14 10:30:25,262 INFO SenderThread:1843 [sender.py:transition_state():613] send defer: 8 +2024-05-14 10:30:25,262 DEBUG SenderThread:1843 [sender.py:send_request():405] send_request: poll_exit +2024-05-14 10:30:25,262 DEBUG HandlerThread:1843 [handler.py:handle_request():158] handle_request: defer +2024-05-14 10:30:25,262 INFO HandlerThread:1843 [handler.py:handle_request_defer():184] handle defer: 8 +2024-05-14 10:30:25,263 DEBUG SenderThread:1843 [sender.py:send_request():405] send_request: defer +2024-05-14 10:30:25,263 INFO SenderThread:1843 [sender.py:send_request_defer():609] handle sender defer: 8 +2024-05-14 10:30:25,263 INFO SenderThread:1843 [job_builder.py:build():432] Attempting to build job artifact +2024-05-14 10:30:25,263 INFO SenderThread:1843 [job_builder.py:_get_source_type():576] no source found +2024-05-14 10:30:25,263 INFO SenderThread:1843 [sender.py:transition_state():613] send defer: 9 +2024-05-14 10:30:25,263 DEBUG HandlerThread:1843 [handler.py:handle_request():158] handle_request: defer +2024-05-14 10:30:25,263 INFO HandlerThread:1843 [handler.py:handle_request_defer():184] handle defer: 9 +2024-05-14 10:30:25,263 DEBUG SenderThread:1843 [sender.py:send_request():405] send_request: defer +2024-05-14 10:30:25,264 INFO SenderThread:1843 [sender.py:send_request_defer():609] handle sender defer: 9 +2024-05-14 10:30:25,264 INFO SenderThread:1843 [dir_watcher.py:finish():358] shutting down directory watcher +2024-05-14 10:30:25,321 INFO SenderThread:1843 [dir_watcher.py:_on_file_modified():288] file/dir modified: /data/cronscript/lm-evaluation-harness/wandb/run-20240514_103002-s6mpkihm/files/output.log +2024-05-14 10:30:25,321 INFO SenderThread:1843 [dir_watcher.py:finish():388] scan: /data/cronscript/lm-evaluation-harness/wandb/run-20240514_103002-s6mpkihm/files +2024-05-14 10:30:25,321 INFO SenderThread:1843 [dir_watcher.py:finish():402] scan save: /data/cronscript/lm-evaluation-harness/wandb/run-20240514_103002-s6mpkihm/files/wandb-summary.json wandb-summary.json +2024-05-14 10:30:25,321 INFO SenderThread:1843 [dir_watcher.py:finish():402] scan save: /data/cronscript/lm-evaluation-harness/wandb/run-20240514_103002-s6mpkihm/files/config.yaml config.yaml +2024-05-14 10:30:25,321 INFO SenderThread:1843 [dir_watcher.py:finish():402] scan save: /data/cronscript/lm-evaluation-harness/wandb/run-20240514_103002-s6mpkihm/files/requirements.txt requirements.txt +2024-05-14 10:30:25,325 INFO SenderThread:1843 [dir_watcher.py:finish():402] scan save: /data/cronscript/lm-evaluation-harness/wandb/run-20240514_103002-s6mpkihm/files/output.log output.log +2024-05-14 10:30:25,326 INFO SenderThread:1843 [dir_watcher.py:finish():402] scan save: /data/cronscript/lm-evaluation-harness/wandb/run-20240514_103002-s6mpkihm/files/wandb-metadata.json wandb-metadata.json +2024-05-14 10:30:25,326 INFO SenderThread:1843 [sender.py:transition_state():613] send defer: 10 +2024-05-14 10:30:25,328 DEBUG HandlerThread:1843 [handler.py:handle_request():158] handle_request: defer +2024-05-14 10:30:25,328 INFO HandlerThread:1843 [handler.py:handle_request_defer():184] handle defer: 10 +2024-05-14 10:30:25,328 DEBUG SenderThread:1843 [sender.py:send_request():405] send_request: defer +2024-05-14 10:30:25,328 INFO SenderThread:1843 [sender.py:send_request_defer():609] handle sender defer: 10 +2024-05-14 10:30:25,328 INFO SenderThread:1843 [file_pusher.py:finish():169] shutting down file pusher +2024-05-14 10:30:25,563 INFO wandb-upload_0:1843 [upload_job.py:push():130] Uploaded file /data/cronscript/lm-evaluation-harness/wandb/run-20240514_103002-s6mpkihm/files/wandb-summary.json +2024-05-14 10:30:25,720 INFO wandb-upload_1:1843 [upload_job.py:push():130] Uploaded file /data/cronscript/lm-evaluation-harness/wandb/run-20240514_103002-s6mpkihm/files/config.yaml +2024-05-14 10:30:25,816 INFO wandb-upload_3:1843 [upload_job.py:push():130] Uploaded file /data/cronscript/lm-evaluation-harness/wandb/run-20240514_103002-s6mpkihm/files/output.log +2024-05-14 10:30:25,831 INFO wandb-upload_2:1843 [upload_job.py:push():130] Uploaded file /data/cronscript/lm-evaluation-harness/wandb/run-20240514_103002-s6mpkihm/files/requirements.txt +2024-05-14 10:30:25,942 DEBUG HandlerThread:1843 [handler.py:handle_request():158] handle_request: poll_exit +2024-05-14 10:30:25,943 DEBUG SenderThread:1843 [sender.py:send_request():405] send_request: poll_exit +2024-05-14 10:30:26,032 INFO Thread-11 (_thread_body):1843 [sender.py:transition_state():613] send defer: 11 +2024-05-14 10:30:26,032 DEBUG HandlerThread:1843 [handler.py:handle_request():158] handle_request: defer +2024-05-14 10:30:26,032 INFO HandlerThread:1843 [handler.py:handle_request_defer():184] handle defer: 11 +2024-05-14 10:30:26,032 DEBUG SenderThread:1843 [sender.py:send_request():405] send_request: defer +2024-05-14 10:30:26,032 INFO SenderThread:1843 [sender.py:send_request_defer():609] handle sender defer: 11 +2024-05-14 10:30:26,032 INFO SenderThread:1843 [file_pusher.py:join():175] waiting for file pusher +2024-05-14 10:30:26,032 INFO SenderThread:1843 [sender.py:transition_state():613] send defer: 12 +2024-05-14 10:30:26,032 DEBUG HandlerThread:1843 [handler.py:handle_request():158] handle_request: defer +2024-05-14 10:30:26,033 INFO HandlerThread:1843 [handler.py:handle_request_defer():184] handle defer: 12 +2024-05-14 10:30:26,033 DEBUG SenderThread:1843 [sender.py:send_request():405] send_request: defer +2024-05-14 10:30:26,033 INFO SenderThread:1843 [sender.py:send_request_defer():609] handle sender defer: 12 +2024-05-14 10:30:26,033 INFO SenderThread:1843 [file_stream.py:finish():601] file stream finish called +2024-05-14 10:30:26,230 INFO SenderThread:1843 [file_stream.py:finish():605] file stream finish is done +2024-05-14 10:30:26,230 INFO SenderThread:1843 [sender.py:transition_state():613] send defer: 13 +2024-05-14 10:30:26,230 DEBUG HandlerThread:1843 [handler.py:handle_request():158] handle_request: defer +2024-05-14 10:30:26,230 INFO HandlerThread:1843 [handler.py:handle_request_defer():184] handle defer: 13 +2024-05-14 10:30:26,230 DEBUG SenderThread:1843 [sender.py:send_request():405] send_request: defer +2024-05-14 10:30:26,230 INFO SenderThread:1843 [sender.py:send_request_defer():609] handle sender defer: 13 +2024-05-14 10:30:26,230 INFO SenderThread:1843 [sender.py:transition_state():613] send defer: 14 +2024-05-14 10:30:26,231 DEBUG HandlerThread:1843 [handler.py:handle_request():158] handle_request: defer +2024-05-14 10:30:26,231 INFO HandlerThread:1843 [handler.py:handle_request_defer():184] handle defer: 14 +2024-05-14 10:30:26,231 DEBUG SenderThread:1843 [sender.py:send():378] send: final +2024-05-14 10:30:26,231 DEBUG SenderThread:1843 [sender.py:send():378] send: footer +2024-05-14 10:30:26,231 DEBUG SenderThread:1843 [sender.py:send_request():405] send_request: defer +2024-05-14 10:30:26,231 INFO SenderThread:1843 [sender.py:send_request_defer():609] handle sender defer: 14 +2024-05-14 10:30:26,231 DEBUG HandlerThread:1843 [handler.py:handle_request():158] handle_request: poll_exit +2024-05-14 10:30:26,232 DEBUG SenderThread:1843 [sender.py:send_request():405] send_request: poll_exit +2024-05-14 10:30:26,232 DEBUG HandlerThread:1843 [handler.py:handle_request():158] handle_request: poll_exit +2024-05-14 10:30:26,232 DEBUG SenderThread:1843 [sender.py:send_request():405] send_request: poll_exit +2024-05-14 10:30:26,232 DEBUG HandlerThread:1843 [handler.py:handle_request():158] handle_request: server_info +2024-05-14 10:30:26,232 DEBUG SenderThread:1843 [sender.py:send_request():405] send_request: server_info +2024-05-14 10:30:26,233 DEBUG HandlerThread:1843 [handler.py:handle_request():158] handle_request: get_summary +2024-05-14 10:30:26,234 DEBUG HandlerThread:1843 [handler.py:handle_request():158] handle_request: sampled_history +2024-05-14 10:30:26,234 DEBUG HandlerThread:1843 [handler.py:handle_request():158] handle_request: internal_messages +2024-05-14 10:30:26,293 INFO MainThread:1843 [wandb_run.py:_footer_history_summary_info():3994] rendering history +2024-05-14 10:30:26,294 INFO MainThread:1843 [wandb_run.py:_footer_history_summary_info():4026] rendering summary +2024-05-14 10:30:26,294 INFO MainThread:1843 [wandb_run.py:_footer_sync_info():3953] logging synced files +2024-05-14 10:30:26,294 DEBUG HandlerThread:1843 [handler.py:handle_request():158] handle_request: shutdown +2024-05-14 10:30:26,294 INFO HandlerThread:1843 [handler.py:finish():882] shutting down handler +2024-05-14 10:30:27,232 INFO WriterThread:1843 [datastore.py:close():296] close: /data/cronscript/lm-evaluation-harness/wandb/run-20240514_103002-s6mpkihm/run-s6mpkihm.wandb +2024-05-14 10:30:27,293 INFO SenderThread:1843 [sender.py:finish():1545] shutting down sender +2024-05-14 10:30:27,293 INFO SenderThread:1843 [file_pusher.py:finish():169] shutting down file pusher +2024-05-14 10:30:27,293 INFO SenderThread:1843 [file_pusher.py:join():175] waiting for file pusher diff --git a/lm-evaluation-harness/wandb/run-20240514_103002-s6mpkihm/logs/debug.log b/lm-evaluation-harness/wandb/run-20240514_103002-s6mpkihm/logs/debug.log new file mode 100644 index 0000000000000000000000000000000000000000..0099cd95cddc18b43d150562b7a35e54e52c6b7a --- /dev/null +++ b/lm-evaluation-harness/wandb/run-20240514_103002-s6mpkihm/logs/debug.log @@ -0,0 +1,29 @@ +2024-05-14 10:30:02,049 INFO MainThread:482 [wandb_setup.py:_flush():76] Current SDK version is 0.17.0 +2024-05-14 10:30:02,049 INFO MainThread:482 [wandb_setup.py:_flush():76] Configure stats pid to 482 +2024-05-14 10:30:02,049 INFO MainThread:482 [wandb_setup.py:_flush():76] Loading settings from /root/.config/wandb/settings +2024-05-14 10:30:02,049 INFO MainThread:482 [wandb_setup.py:_flush():76] Loading settings from /data/cronscript/lm-evaluation-harness/wandb/settings +2024-05-14 10:30:02,049 INFO MainThread:482 [wandb_setup.py:_flush():76] Loading settings from environment variables: {} +2024-05-14 10:30:02,049 INFO MainThread:482 [wandb_setup.py:_flush():76] Applying setup settings: {'_disable_service': False} +2024-05-14 10:30:02,049 WARNING MainThread:482 [wandb_setup.py:_flush():76] Could not find program at -m lm_eval.__main__ +2024-05-14 10:30:02,049 INFO MainThread:482 [wandb_setup.py:_flush():76] Inferring run settings from compute environment: {'program_relpath': None, 'program': '-m lm_eval.__main__'} +2024-05-14 10:30:02,049 INFO MainThread:482 [wandb_setup.py:_flush():76] Applying login settings: {} +2024-05-14 10:30:02,050 INFO MainThread:482 [wandb_init.py:_log_setup():520] Logging user logs to /data/cronscript/lm-evaluation-harness/wandb/run-20240514_103002-s6mpkihm/logs/debug.log +2024-05-14 10:30:02,050 INFO MainThread:482 [wandb_init.py:_log_setup():521] Logging internal logs to /data/cronscript/lm-evaluation-harness/wandb/run-20240514_103002-s6mpkihm/logs/debug-internal.log +2024-05-14 10:30:02,050 INFO MainThread:482 [wandb_init.py:init():560] calling init triggers +2024-05-14 10:30:02,050 INFO MainThread:482 [wandb_init.py:init():567] wandb.init called with sweep_config: {} +config: {} +2024-05-14 10:30:02,050 INFO MainThread:482 [wandb_init.py:init():610] starting backend +2024-05-14 10:30:02,050 INFO MainThread:482 [wandb_init.py:init():614] setting up manager +2024-05-14 10:30:02,051 INFO MainThread:482 [backend.py:_multiprocessing_setup():105] multiprocessing start_methods=fork,spawn,forkserver, using: spawn +2024-05-14 10:30:02,052 INFO MainThread:482 [wandb_init.py:init():622] backend started and connected +2024-05-14 10:30:02,054 INFO MainThread:482 [wandb_init.py:init():711] updated telemetry +2024-05-14 10:30:02,065 INFO MainThread:482 [wandb_init.py:init():744] communicating run to backend with 90.0 second timeout +2024-05-14 10:30:02,301 INFO MainThread:482 [wandb_run.py:_on_init():2396] communicating current version +2024-05-14 10:30:02,378 INFO MainThread:482 [wandb_run.py:_on_init():2405] got version response +2024-05-14 10:30:02,378 INFO MainThread:482 [wandb_init.py:init():795] starting run threads in backend +2024-05-14 10:30:02,596 INFO MainThread:482 [wandb_run.py:_console_start():2374] atexit reg +2024-05-14 10:30:02,597 INFO MainThread:482 [wandb_run.py:_redirect():2229] redirect: wrap_raw +2024-05-14 10:30:02,597 INFO MainThread:482 [wandb_run.py:_redirect():2294] Wrapping output streams. +2024-05-14 10:30:02,597 INFO MainThread:482 [wandb_run.py:_redirect():2319] Redirects installed. +2024-05-14 10:30:02,598 INFO MainThread:482 [wandb_init.py:init():838] run started, returning control to user process +2024-05-14 10:30:27,295 WARNING MsgRouterThr:482 [router.py:message_loop():77] message_loop has been closed diff --git a/lm-evaluation-harness/wandb/run-20240514_103049-lsh767q1/files/config.yaml b/lm-evaluation-harness/wandb/run-20240514_103049-lsh767q1/files/config.yaml new file mode 100644 index 0000000000000000000000000000000000000000..a0ad52cfc46bc0b2c4d98a16521a8de39500bdc2 --- /dev/null +++ b/lm-evaluation-harness/wandb/run-20240514_103049-lsh767q1/files/config.yaml @@ -0,0 +1,43 @@ +wandb_version: 1 + +_wandb: + desc: null + value: + python_version: 3.10.12 + cli_version: 0.17.0 + framework: huggingface + huggingface_version: 4.40.2 + is_jupyter_run: false + is_kaggle_kernel: false + start_time: 1715682649 + t: + 1: + - 1 + - 5 + - 11 + - 49 + - 51 + - 53 + - 55 + - 71 + - 98 + - 100 + 2: + - 1 + - 5 + - 11 + - 49 + - 51 + - 53 + - 55 + - 71 + - 98 + - 100 + 3: + - 23 + 4: 3.10.12 + 5: 0.17.0 + 6: 4.40.2 + 8: + - 5 + 13: linux-x86_64 diff --git a/lm-evaluation-harness/wandb/run-20240514_103049-lsh767q1/files/output.log b/lm-evaluation-harness/wandb/run-20240514_103049-lsh767q1/files/output.log new file mode 100644 index 0000000000000000000000000000000000000000..3eac565516dd522c20cf43676fec63f3d54d1012 --- /dev/null +++ b/lm-evaluation-harness/wandb/run-20240514_103049-lsh767q1/files/output.log @@ -0,0 +1,42 @@ + +2024-05-14:10:30:50,211 INFO [__main__.py:251] Verbosity set to INFO +2024-05-14:10:30:54,793 INFO [__main__.py:335] Selected Tasks: ['indiccopa-hi'] +2024-05-14:10:30:54,795 INFO [evaluator.py:131] Setting random seed to 0 | Setting numpy seed to 1234 | Setting torch manual seed to 1234 +2024-05-14:10:30:54,795 INFO [evaluator.py:177] Initializing hf model, with arguments: {'pretrained': '/data/cronscript/ckpts//hf_ckpt//global_step120'} +Traceback (most recent call last): + File "/usr/local/lib/python3.10/dist-packages/transformers/utils/hub.py", line 398, in cached_file + resolved_file = hf_hub_download( + File "/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_validators.py", line 106, in _inner_fn + validate_repo_id(arg_value) + File "/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_validators.py", line 154, in validate_repo_id + raise HFValidationError( +huggingface_hub.errors.HFValidationError: Repo id must be in the form 'repo_name' or 'namespace/repo_name': '/data/cronscript/ckpts//hf_ckpt//global_step120'. Use `repo_type` argument if needed. +The above exception was the direct cause of the following exception: +Traceback (most recent call last): + File "/usr/lib/python3.10/runpy.py", line 196, in _run_module_as_main + return _run_code(code, main_globals, None, + File "/usr/lib/python3.10/runpy.py", line 86, in _run_code + exec(code, run_globals) + File "/data/cronscript/lm-evaluation-harness/lm_eval/__main__.py", line 417, in + cli_evaluate() + File "/data/cronscript/lm-evaluation-harness/lm_eval/__main__.py", line 341, in cli_evaluate + results = evaluator.simple_evaluate( + File "/data/cronscript/lm-evaluation-harness/lm_eval/utils.py", line 288, in _wrapper + return fn(*args, **kwargs) + File "/data/cronscript/lm-evaluation-harness/lm_eval/evaluator.py", line 180, in simple_evaluate + lm = lm_eval.api.registry.get_model(model).create_from_arg_string( + File "/data/cronscript/lm-evaluation-harness/lm_eval/api/model.py", line 134, in create_from_arg_string + return cls(**args, **args2) + File "/data/cronscript/lm-evaluation-harness/lm_eval/models/huggingface.py", line 190, in __init__ + self._get_config( + File "/data/cronscript/lm-evaluation-harness/lm_eval/models/huggingface.py", line 471, in _get_config + self._config = transformers.AutoConfig.from_pretrained( + File "/usr/local/lib/python3.10/dist-packages/transformers/models/auto/configuration_auto.py", line 928, in from_pretrained + config_dict, unused_kwargs = PretrainedConfig.get_config_dict(pretrained_model_name_or_path, **kwargs) + File "/usr/local/lib/python3.10/dist-packages/transformers/configuration_utils.py", line 631, in get_config_dict + config_dict, kwargs = cls._get_config_dict(pretrained_model_name_or_path, **kwargs) + File "/usr/local/lib/python3.10/dist-packages/transformers/configuration_utils.py", line 686, in _get_config_dict + resolved_config_file = cached_file( + File "/usr/local/lib/python3.10/dist-packages/transformers/utils/hub.py", line 462, in cached_file + raise EnvironmentError( +OSError: Incorrect path_or_model_id: '/data/cronscript/ckpts//hf_ckpt//global_step120'. Please provide either the path to a local folder or the repo_id of a model on the Hub. \ No newline at end of file diff --git a/lm-evaluation-harness/wandb/run-20240514_103049-lsh767q1/files/requirements.txt b/lm-evaluation-harness/wandb/run-20240514_103049-lsh767q1/files/requirements.txt new file mode 100644 index 0000000000000000000000000000000000000000..d63edd781bd5bdbb7f67523ac1ba9f0f1ed392dc --- /dev/null +++ b/lm-evaluation-harness/wandb/run-20240514_103049-lsh767q1/files/requirements.txt @@ -0,0 +1,163 @@ +DataProperty==1.0.1 +GitPython==3.1.43 +Jinja2==3.1.3 +Markdown==3.6 +MarkupSafe==2.1.5 +Pillow-SIMD==7.0.0.post3 +PyYAML==6.0 +Werkzeug==3.0.2 +absl-py==2.1.0 +accelerate==0.30.1 +aiohttp==3.9.4 +aiosignal==1.3.1 +antlr4-python3-runtime==4.9.3 +anyio==4.3.0 +async-timeout==4.0.3 +attrs==23.2.0 +av==9.2.0 +cachetools==5.3.3 +certifi==2024.2.2 +cffi==1.15.1 +cfgv==3.4.0 +chardet==5.2.0 +charset-normalizer==3.3.2 +click==8.1.7 +cmake==3.29.2 +colorama==0.4.6 +datasets==2.19.1 +deepspeed==0.12.4+hpu.synapse.v1.15.1 +dill==0.3.8 +distlib==0.3.8 +distro==1.9.0 +docker-pycreds==0.4.0 +einops==0.8.0 +evaluate==0.4.2 +exceptiongroup==1.2.0 +expecttest==0.2.1 +filelock==3.13.4 +frozenlist==1.4.1 +fsspec==2024.3.1 +gitdb==4.0.11 +google-auth-oauthlib==0.4.6 +google-auth==2.29.0 +grpcio==1.62.1 +h11==0.14.0 +habana-media-loader==1.15.1.15 +habana-pyhlml==1.15.1.15 +habana-torch-dataloader==1.15.1.15 +habana-torch-plugin==1.15.1.15 +habana_gpu_migration==1.15.1.15 +habana_quantization_toolkit==1.15.1.15 +hjson==3.1.0 +httpcore==1.0.5 +httpx==0.27.0 +huggingface-hub==0.23.0 +identify==2.5.35 +idna==3.7 +importlib_resources==6.4.0 +iniconfig==2.0.0 +joblib==1.4.2 +jsonlines==4.0.0 +lightning-habana==1.4.0 +lightning-utilities==0.11.2 +lightning==2.2.0.post0 +lm_eval==0.3.0 +lm_eval==0.4.2 +lm_eval==0.4.2 +lm_eval==0.4.2 +mbstrdecoder==1.1.3 +more-itertools==10.2.0 +mpi4py==3.1.4 +mpmath==1.3.0 +multidict==6.0.5 +multiprocess==0.70.16 +networkx==3.3 +ninja==1.11.1.1 +nltk==3.8.1 +nodeenv==1.8.0 +numexpr==2.10.0 +numpy==1.23.5 +oauthlib==3.2.2 +omegaconf==2.3.0 +openai==1.29.0 +packaging==24.0 +pandas==2.0.1 +pathspec==0.12.1 +pathvalidate==3.2.0 +peft==0.10.0 +perfetto==0.7.0 +pip==22.0.2 +pip==23.3.1 +platformdirs==4.2.0 +pluggy==1.4.0 +portalocker==2.8.2 +pre-commit==3.3.3 +protobuf==3.20.3 +psutil==5.9.8 +py-cpuinfo==9.0.0 +pyarrow-hotfix==0.6 +pyarrow==16.0.0 +pyasn1==0.6.0 +pyasn1_modules==0.4.0 +pybind11==2.10.4 +pycountry==23.12.11 +pycparser==2.22 +pydantic==1.10.13 +pynvml==8.0.4 +pytablewriter==1.2.0 +pytest==8.1.1 +python-dateutil==2.9.0.post0 +pytorch-lightning==2.2.2 +pytz==2024.1 +regex==2023.5.5 +requests-oauthlib==2.0.0 +requests==2.31.0 +rouge_score==0.1.2 +rsa==4.9 +sacrebleu==1.5.0 +safetensors==0.4.3 +scikit-learn==1.4.2 +scipy==1.13.0 +sentencepiece==0.2.0 +sentry-sdk==2.1.1 +setproctitle==1.3.3 +setuptools==59.6.0 +setuptools==69.5.1 +six==1.16.0 +smmap==5.0.1 +sniffio==1.3.1 +sqlitedict==2.1.0 +symengine==0.11.0 +sympy==1.12 +tabledata==1.3.3 +tcolorpy==0.1.6 +tdqm==0.0.1 +tensorboard-data-server==0.6.1 +tensorboard-plugin-wit==1.8.1 +tensorboard==2.11.2 +threadpoolctl==3.5.0 +tokenizers==0.19.1 +tomli==2.0.1 +torch==2.2.0a0+git8964477 +torch_tb_profiler==0.4.0 +torchaudio==2.2.0+08901ad +torchdata==0.7.1+5e6f7b7 +torchmetrics==1.3.2 +torchtext==0.17.0+400da5c +torchvision==0.17.0+b2383d4 +tqdm-multiprocess==0.0.11 +tqdm==4.66.2 +transformers==4.40.2 +typepy==1.3.2 +typing_extensions==4.11.0 +tzdata==2024.1 +urllib3==1.26.18 +virtualenv==20.25.1 +wandb==0.17.0 +wheel==0.37.1 +wheel==0.43.0 +word2number==1.1 +xxhash==3.4.1 +yamllint==1.35.1 +yarl==1.9.4 +zstandard==0.22.0 \ No newline at end of file diff --git a/lm-evaluation-harness/wandb/run-20240514_103049-lsh767q1/files/wandb-metadata.json b/lm-evaluation-harness/wandb/run-20240514_103049-lsh767q1/files/wandb-metadata.json new file mode 100644 index 0000000000000000000000000000000000000000..40e2f655bccb6a8ece335acacc81b7107746fd96 --- /dev/null +++ b/lm-evaluation-harness/wandb/run-20240514_103049-lsh767q1/files/wandb-metadata.json @@ -0,0 +1,810 @@ +{ + "os": "Linux-5.15.0-92-generic-x86_64-with-glibc2.35", + "python": "3.10.12", + "heartbeatAt": "2024-05-14T10:30:50.080441", + "startedAt": "2024-05-14T10:30:49.689603", + "docker": null, + "cuda": null, + "args": [ + "--model", + "hf", + "--model_args", + "pretrained=/data/cronscript/ckpts//hf_ckpt//global_step120", + "--tasks", + "indiccopa-hi", + "--batch_size", + "auto", + "--wandb_args", + "project=bharatgpt" + ], + "state": "running", + "program": "-m lm_eval.__main__", + "codePathLocal": null, + "git": { + "remote": "https://github.com/EleutherAI/lm-evaluation-harness", + "commit": null + }, + "email": null, + "root": "/data/cronscript/lm-evaluation-harness", + "host": "vizzhy-150-3", + "username": "root", + "executable": "/usr/bin/python3", + "cpu_count": 76, + "cpu_count_logical": 152, + "cpu_freq": { + "current": 3397.59875, + "min": 800.0, + "max": 3400.0 + }, + "cpu_freq_per_core": [ + { + "current": 3300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3300.003, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.003, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + } + ], + "disk": { + "/": { + "total": 866.4415092468262, + "used": 76.92232131958008 + } + }, + "memory": { + "total": 1007.5000267028809 + } +} diff --git a/lm-evaluation-harness/wandb/run-20240514_103049-lsh767q1/files/wandb-summary.json b/lm-evaluation-harness/wandb/run-20240514_103049-lsh767q1/files/wandb-summary.json new file mode 100644 index 0000000000000000000000000000000000000000..e682bae6b5eaeba8295fd0fffdc51474a259249e --- /dev/null +++ b/lm-evaluation-harness/wandb/run-20240514_103049-lsh767q1/files/wandb-summary.json @@ -0,0 +1 @@ +{"_wandb": {"runtime": 5}} \ No newline at end of file diff --git a/lm-evaluation-harness/wandb/run-20240514_103049-lsh767q1/logs/debug-internal.log b/lm-evaluation-harness/wandb/run-20240514_103049-lsh767q1/logs/debug-internal.log new file mode 100644 index 0000000000000000000000000000000000000000..49edc96737dd2d65c1d42dad4adc2282bcb80328 --- /dev/null +++ b/lm-evaluation-harness/wandb/run-20240514_103049-lsh767q1/logs/debug-internal.log @@ -0,0 +1,179 @@ +2024-05-14 10:30:49,702 INFO StreamThr :7281 [internal.py:wandb_internal():85] W&B internal server running at pid: 7281, started at: 2024-05-14 10:30:49.701782 +2024-05-14 10:30:49,704 DEBUG HandlerThread:7281 [handler.py:handle_request():158] handle_request: status +2024-05-14 10:30:49,705 INFO WriterThread:7281 [datastore.py:open_for_write():87] open: /data/cronscript/lm-evaluation-harness/wandb/run-20240514_103049-lsh767q1/run-lsh767q1.wandb +2024-05-14 10:30:49,706 DEBUG SenderThread:7281 [sender.py:send():378] send: header +2024-05-14 10:30:49,715 DEBUG SenderThread:7281 [sender.py:send():378] send: run +2024-05-14 10:30:49,924 INFO SenderThread:7281 [dir_watcher.py:__init__():211] watching files in: /data/cronscript/lm-evaluation-harness/wandb/run-20240514_103049-lsh767q1/files +2024-05-14 10:30:49,924 INFO SenderThread:7281 [sender.py:_start_run_threads():1123] run started: lsh767q1 with start time 1715682649.701201 +2024-05-14 10:30:49,930 DEBUG HandlerThread:7281 [handler.py:handle_request():158] handle_request: check_version +2024-05-14 10:30:49,930 DEBUG SenderThread:7281 [sender.py:send_request():405] send_request: check_version +2024-05-14 10:30:50,012 DEBUG HandlerThread:7281 [handler.py:handle_request():158] handle_request: run_start +2024-05-14 10:30:50,014 DEBUG HandlerThread:7281 [system_info.py:__init__():26] System info init +2024-05-14 10:30:50,014 DEBUG HandlerThread:7281 [system_info.py:__init__():41] System info init done +2024-05-14 10:30:50,014 INFO HandlerThread:7281 [system_monitor.py:start():194] Starting system monitor +2024-05-14 10:30:50,014 INFO SystemMonitor:7281 [system_monitor.py:_start():158] Starting system asset monitoring threads +2024-05-14 10:30:50,014 INFO HandlerThread:7281 [system_monitor.py:probe():214] Collecting system info +2024-05-14 10:30:50,014 INFO SystemMonitor:7281 [interfaces.py:start():188] Started cpu monitoring +2024-05-14 10:30:50,015 INFO SystemMonitor:7281 [interfaces.py:start():188] Started disk monitoring +2024-05-14 10:30:50,016 INFO SystemMonitor:7281 [interfaces.py:start():188] Started memory monitoring +2024-05-14 10:30:50,016 INFO SystemMonitor:7281 [interfaces.py:start():188] Started network monitoring +2024-05-14 10:30:50,080 DEBUG HandlerThread:7281 [system_info.py:probe():150] Probing system +2024-05-14 10:30:50,089 DEBUG HandlerThread:7281 [system_info.py:_probe_git():135] Probing git +2024-05-14 10:30:50,108 ERROR HandlerThread:7281 [gitlib.py:root():92] git root error: Cmd('git') failed due to: exit code(128) + cmdline: git rev-parse --show-toplevel + stderr: 'fatal: detected dubious ownership in repository at '/data/cronscript/lm-evaluation-harness' +To add an exception for this directory, call: + + git config --global --add safe.directory /data/cronscript/lm-evaluation-harness' +2024-05-14 10:30:50,108 DEBUG HandlerThread:7281 [system_info.py:_probe_git():143] Probing git done +2024-05-14 10:30:50,109 DEBUG HandlerThread:7281 [system_info.py:probe():198] Probing system done +2024-05-14 10:30:50,109 DEBUG HandlerThread:7281 [system_monitor.py:probe():223] {'os': 'Linux-5.15.0-92-generic-x86_64-with-glibc2.35', 'python': '3.10.12', 'heartbeatAt': '2024-05-14T10:30:50.080441', 'startedAt': '2024-05-14T10:30:49.689603', 'docker': None, 'cuda': None, 'args': ('--model', 'hf', '--model_args', 'pretrained=/data/cronscript/ckpts//hf_ckpt//global_step120', '--tasks', 'indiccopa-hi', '--batch_size', 'auto', '--wandb_args', 'project=bharatgpt'), 'state': 'running', 'program': '-m lm_eval.__main__', 'codePathLocal': None, 'git': {'remote': 'https://github.com/EleutherAI/lm-evaluation-harness', 'commit': None}, 'email': None, 'root': '/data/cronscript/lm-evaluation-harness', 'host': 'vizzhy-150-3', 'username': 'root', 'executable': '/usr/bin/python3', 'cpu_count': 76, 'cpu_count_logical': 152, 'cpu_freq': {'current': 3397.59875, 'min': 800.0, 'max': 3400.0}, 'cpu_freq_per_core': [{'current': 3300.0, 'min': 800.0, 'max': 3400.0}, {'current': 3300.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3300.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3300.003, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.003, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}], 'disk': {'/': {'total': 866.4415092468262, 'used': 76.92232131958008}}, 'memory': {'total': 1007.5000267028809}} +2024-05-14 10:30:50,109 INFO HandlerThread:7281 [system_monitor.py:probe():224] Finished collecting system info +2024-05-14 10:30:50,109 INFO HandlerThread:7281 [system_monitor.py:probe():227] Publishing system info +2024-05-14 10:30:50,110 INFO HandlerThread:7281 [system_monitor.py:probe():229] Finished publishing system info +2024-05-14 10:30:50,113 DEBUG SenderThread:7281 [sender.py:send():378] send: files +2024-05-14 10:30:50,113 INFO SenderThread:7281 [sender.py:_save_file():1389] saving file wandb-metadata.json with policy now +2024-05-14 10:30:50,208 DEBUG HandlerThread:7281 [handler.py:handle_request():158] handle_request: python_packages +2024-05-14 10:30:50,208 DEBUG SenderThread:7281 [sender.py:send_request():405] send_request: python_packages +2024-05-14 10:30:50,208 DEBUG HandlerThread:7281 [handler.py:handle_request():158] handle_request: stop_status +2024-05-14 10:30:50,209 DEBUG SenderThread:7281 [sender.py:send_request():405] send_request: stop_status +2024-05-14 10:30:50,410 DEBUG SenderThread:7281 [sender.py:send():378] send: telemetry +2024-05-14 10:30:50,677 INFO wandb-upload_0:7281 [upload_job.py:push():130] Uploaded file /tmp/tmp87d4y4upwandb/6j4sd0kd-wandb-metadata.json +2024-05-14 10:30:50,925 INFO Thread-12 :7281 [dir_watcher.py:_on_file_created():271] file/dir created: /data/cronscript/lm-evaluation-harness/wandb/run-20240514_103049-lsh767q1/files/output.log +2024-05-14 10:30:50,926 INFO Thread-12 :7281 [dir_watcher.py:_on_file_created():271] file/dir created: /data/cronscript/lm-evaluation-harness/wandb/run-20240514_103049-lsh767q1/files/requirements.txt +2024-05-14 10:30:50,926 INFO Thread-12 :7281 [dir_watcher.py:_on_file_created():271] file/dir created: /data/cronscript/lm-evaluation-harness/wandb/run-20240514_103049-lsh767q1/files/wandb-metadata.json +2024-05-14 10:30:52,926 INFO Thread-12 :7281 [dir_watcher.py:_on_file_modified():288] file/dir modified: /data/cronscript/lm-evaluation-harness/wandb/run-20240514_103049-lsh767q1/files/output.log +2024-05-14 10:30:54,794 DEBUG HandlerThread:7281 [handler.py:handle_request():158] handle_request: status_report +2024-05-14 10:30:55,867 DEBUG SenderThread:7281 [sender.py:send():378] send: exit +2024-05-14 10:30:55,867 INFO SenderThread:7281 [sender.py:send_exit():585] handling exit code: 1 +2024-05-14 10:30:55,868 INFO SenderThread:7281 [sender.py:send_exit():587] handling runtime: 5 +2024-05-14 10:30:55,869 INFO SenderThread:7281 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end +2024-05-14 10:30:55,869 INFO SenderThread:7281 [sender.py:send_exit():593] send defer +2024-05-14 10:30:55,869 DEBUG HandlerThread:7281 [handler.py:handle_request():158] handle_request: defer +2024-05-14 10:30:55,869 INFO HandlerThread:7281 [handler.py:handle_request_defer():184] handle defer: 0 +2024-05-14 10:30:55,869 DEBUG SenderThread:7281 [sender.py:send_request():405] send_request: defer +2024-05-14 10:30:55,869 INFO SenderThread:7281 [sender.py:send_request_defer():609] handle sender defer: 0 +2024-05-14 10:30:55,869 INFO SenderThread:7281 [sender.py:transition_state():613] send defer: 1 +2024-05-14 10:30:55,869 DEBUG HandlerThread:7281 [handler.py:handle_request():158] handle_request: defer +2024-05-14 10:30:55,869 INFO HandlerThread:7281 [handler.py:handle_request_defer():184] handle defer: 1 +2024-05-14 10:30:55,869 DEBUG SenderThread:7281 [sender.py:send_request():405] send_request: defer +2024-05-14 10:30:55,869 INFO SenderThread:7281 [sender.py:send_request_defer():609] handle sender defer: 1 +2024-05-14 10:30:55,869 INFO SenderThread:7281 [sender.py:transition_state():613] send defer: 2 +2024-05-14 10:30:55,869 DEBUG HandlerThread:7281 [handler.py:handle_request():158] handle_request: defer +2024-05-14 10:30:55,869 INFO HandlerThread:7281 [handler.py:handle_request_defer():184] handle defer: 2 +2024-05-14 10:30:55,870 INFO HandlerThread:7281 [system_monitor.py:finish():203] Stopping system monitor +2024-05-14 10:30:55,870 DEBUG SystemMonitor:7281 [system_monitor.py:_start():172] Starting system metrics aggregation loop +2024-05-14 10:30:55,870 DEBUG SystemMonitor:7281 [system_monitor.py:_start():179] Finished system metrics aggregation loop +2024-05-14 10:30:55,870 DEBUG SystemMonitor:7281 [system_monitor.py:_start():183] Publishing last batch of metrics +2024-05-14 10:30:55,870 INFO HandlerThread:7281 [interfaces.py:finish():200] Joined cpu monitor +2024-05-14 10:30:55,871 INFO HandlerThread:7281 [interfaces.py:finish():200] Joined disk monitor +2024-05-14 10:30:55,871 INFO HandlerThread:7281 [interfaces.py:finish():200] Joined memory monitor +2024-05-14 10:30:55,871 INFO HandlerThread:7281 [interfaces.py:finish():200] Joined network monitor +2024-05-14 10:30:55,871 DEBUG SenderThread:7281 [sender.py:send_request():405] send_request: defer +2024-05-14 10:30:55,871 INFO SenderThread:7281 [sender.py:send_request_defer():609] handle sender defer: 2 +2024-05-14 10:30:55,871 INFO SenderThread:7281 [sender.py:transition_state():613] send defer: 3 +2024-05-14 10:30:55,871 DEBUG SenderThread:7281 [sender.py:send():378] send: stats +2024-05-14 10:30:55,871 DEBUG HandlerThread:7281 [handler.py:handle_request():158] handle_request: defer +2024-05-14 10:30:55,871 INFO HandlerThread:7281 [handler.py:handle_request_defer():184] handle defer: 3 +2024-05-14 10:30:55,871 DEBUG SenderThread:7281 [sender.py:send_request():405] send_request: defer +2024-05-14 10:30:55,871 INFO SenderThread:7281 [sender.py:send_request_defer():609] handle sender defer: 3 +2024-05-14 10:30:55,871 INFO SenderThread:7281 [sender.py:transition_state():613] send defer: 4 +2024-05-14 10:30:55,872 DEBUG HandlerThread:7281 [handler.py:handle_request():158] handle_request: defer +2024-05-14 10:30:55,872 INFO HandlerThread:7281 [handler.py:handle_request_defer():184] handle defer: 4 +2024-05-14 10:30:55,872 DEBUG SenderThread:7281 [sender.py:send_request():405] send_request: defer +2024-05-14 10:30:55,872 INFO SenderThread:7281 [sender.py:send_request_defer():609] handle sender defer: 4 +2024-05-14 10:30:55,872 INFO SenderThread:7281 [sender.py:transition_state():613] send defer: 5 +2024-05-14 10:30:55,872 DEBUG HandlerThread:7281 [handler.py:handle_request():158] handle_request: defer +2024-05-14 10:30:55,872 INFO HandlerThread:7281 [handler.py:handle_request_defer():184] handle defer: 5 +2024-05-14 10:30:55,872 DEBUG SenderThread:7281 [sender.py:send():378] send: summary +2024-05-14 10:30:55,873 INFO SenderThread:7281 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end +2024-05-14 10:30:55,873 DEBUG SenderThread:7281 [sender.py:send_request():405] send_request: defer +2024-05-14 10:30:55,873 INFO SenderThread:7281 [sender.py:send_request_defer():609] handle sender defer: 5 +2024-05-14 10:30:55,873 INFO SenderThread:7281 [sender.py:transition_state():613] send defer: 6 +2024-05-14 10:30:55,873 DEBUG HandlerThread:7281 [handler.py:handle_request():158] handle_request: defer +2024-05-14 10:30:55,873 INFO HandlerThread:7281 [handler.py:handle_request_defer():184] handle defer: 6 +2024-05-14 10:30:55,873 DEBUG SenderThread:7281 [sender.py:send_request():405] send_request: defer +2024-05-14 10:30:55,873 INFO SenderThread:7281 [sender.py:send_request_defer():609] handle sender defer: 6 +2024-05-14 10:30:55,875 DEBUG HandlerThread:7281 [handler.py:handle_request():158] handle_request: status_report +2024-05-14 10:30:55,928 INFO Thread-12 :7281 [dir_watcher.py:_on_file_created():271] file/dir created: /data/cronscript/lm-evaluation-harness/wandb/run-20240514_103049-lsh767q1/files/wandb-summary.json +2024-05-14 10:30:55,962 INFO SenderThread:7281 [sender.py:transition_state():613] send defer: 7 +2024-05-14 10:30:55,962 DEBUG HandlerThread:7281 [handler.py:handle_request():158] handle_request: defer +2024-05-14 10:30:55,962 INFO HandlerThread:7281 [handler.py:handle_request_defer():184] handle defer: 7 +2024-05-14 10:30:55,962 DEBUG SenderThread:7281 [sender.py:send_request():405] send_request: defer +2024-05-14 10:30:55,962 INFO SenderThread:7281 [sender.py:send_request_defer():609] handle sender defer: 7 +2024-05-14 10:30:56,430 INFO SenderThread:7281 [sender.py:transition_state():613] send defer: 8 +2024-05-14 10:30:56,430 DEBUG HandlerThread:7281 [handler.py:handle_request():158] handle_request: defer +2024-05-14 10:30:56,430 INFO HandlerThread:7281 [handler.py:handle_request_defer():184] handle defer: 8 +2024-05-14 10:30:56,430 DEBUG SenderThread:7281 [sender.py:send_request():405] send_request: defer +2024-05-14 10:30:56,430 INFO SenderThread:7281 [sender.py:send_request_defer():609] handle sender defer: 8 +2024-05-14 10:30:56,430 INFO SenderThread:7281 [job_builder.py:build():432] Attempting to build job artifact +2024-05-14 10:30:56,431 INFO SenderThread:7281 [job_builder.py:_get_source_type():576] no source found +2024-05-14 10:30:56,431 INFO SenderThread:7281 [sender.py:transition_state():613] send defer: 9 +2024-05-14 10:30:56,431 DEBUG HandlerThread:7281 [handler.py:handle_request():158] handle_request: defer +2024-05-14 10:30:56,431 INFO HandlerThread:7281 [handler.py:handle_request_defer():184] handle defer: 9 +2024-05-14 10:30:56,431 DEBUG SenderThread:7281 [sender.py:send_request():405] send_request: defer +2024-05-14 10:30:56,431 INFO SenderThread:7281 [sender.py:send_request_defer():609] handle sender defer: 9 +2024-05-14 10:30:56,431 INFO SenderThread:7281 [dir_watcher.py:finish():358] shutting down directory watcher +2024-05-14 10:30:56,868 DEBUG HandlerThread:7281 [handler.py:handle_request():158] handle_request: poll_exit +2024-05-14 10:30:56,929 INFO SenderThread:7281 [dir_watcher.py:_on_file_modified():288] file/dir modified: /data/cronscript/lm-evaluation-harness/wandb/run-20240514_103049-lsh767q1/files/output.log +2024-05-14 10:30:56,929 INFO SenderThread:7281 [dir_watcher.py:_on_file_modified():288] file/dir modified: /data/cronscript/lm-evaluation-harness/wandb/run-20240514_103049-lsh767q1/files/config.yaml +2024-05-14 10:30:56,930 INFO SenderThread:7281 [dir_watcher.py:finish():388] scan: /data/cronscript/lm-evaluation-harness/wandb/run-20240514_103049-lsh767q1/files +2024-05-14 10:30:56,930 INFO SenderThread:7281 [dir_watcher.py:finish():402] scan save: /data/cronscript/lm-evaluation-harness/wandb/run-20240514_103049-lsh767q1/files/wandb-metadata.json wandb-metadata.json +2024-05-14 10:30:56,930 INFO SenderThread:7281 [dir_watcher.py:finish():402] scan save: /data/cronscript/lm-evaluation-harness/wandb/run-20240514_103049-lsh767q1/files/output.log output.log +2024-05-14 10:30:56,930 INFO SenderThread:7281 [dir_watcher.py:finish():402] scan save: /data/cronscript/lm-evaluation-harness/wandb/run-20240514_103049-lsh767q1/files/requirements.txt requirements.txt +2024-05-14 10:30:56,930 INFO SenderThread:7281 [dir_watcher.py:finish():402] scan save: /data/cronscript/lm-evaluation-harness/wandb/run-20240514_103049-lsh767q1/files/config.yaml config.yaml +2024-05-14 10:30:56,930 INFO SenderThread:7281 [dir_watcher.py:finish():402] scan save: /data/cronscript/lm-evaluation-harness/wandb/run-20240514_103049-lsh767q1/files/wandb-summary.json wandb-summary.json +2024-05-14 10:30:56,933 INFO SenderThread:7281 [sender.py:transition_state():613] send defer: 10 +2024-05-14 10:30:56,933 DEBUG SenderThread:7281 [sender.py:send_request():405] send_request: poll_exit +2024-05-14 10:30:56,933 DEBUG HandlerThread:7281 [handler.py:handle_request():158] handle_request: defer +2024-05-14 10:30:56,933 INFO HandlerThread:7281 [handler.py:handle_request_defer():184] handle defer: 10 +2024-05-14 10:30:56,936 DEBUG SenderThread:7281 [sender.py:send_request():405] send_request: defer +2024-05-14 10:30:56,937 INFO SenderThread:7281 [sender.py:send_request_defer():609] handle sender defer: 10 +2024-05-14 10:30:56,937 INFO SenderThread:7281 [file_pusher.py:finish():169] shutting down file pusher +2024-05-14 10:30:57,171 INFO wandb-upload_1:7281 [upload_job.py:push():130] Uploaded file /data/cronscript/lm-evaluation-harness/wandb/run-20240514_103049-lsh767q1/files/output.log +2024-05-14 10:30:57,332 INFO wandb-upload_0:7281 [upload_job.py:push():130] Uploaded file /data/cronscript/lm-evaluation-harness/wandb/run-20240514_103049-lsh767q1/files/requirements.txt +2024-05-14 10:30:57,407 INFO wandb-upload_3:7281 [upload_job.py:push():130] Uploaded file /data/cronscript/lm-evaluation-harness/wandb/run-20240514_103049-lsh767q1/files/wandb-summary.json +2024-05-14 10:30:57,418 INFO wandb-upload_2:7281 [upload_job.py:push():130] Uploaded file /data/cronscript/lm-evaluation-harness/wandb/run-20240514_103049-lsh767q1/files/config.yaml +2024-05-14 10:30:57,618 INFO Thread-11 (_thread_body):7281 [sender.py:transition_state():613] send defer: 11 +2024-05-14 10:30:57,619 DEBUG HandlerThread:7281 [handler.py:handle_request():158] handle_request: defer +2024-05-14 10:30:57,619 INFO HandlerThread:7281 [handler.py:handle_request_defer():184] handle defer: 11 +2024-05-14 10:30:57,620 DEBUG SenderThread:7281 [sender.py:send_request():405] send_request: defer +2024-05-14 10:30:57,620 INFO SenderThread:7281 [sender.py:send_request_defer():609] handle sender defer: 11 +2024-05-14 10:30:57,620 INFO SenderThread:7281 [file_pusher.py:join():175] waiting for file pusher +2024-05-14 10:30:57,620 INFO SenderThread:7281 [sender.py:transition_state():613] send defer: 12 +2024-05-14 10:30:57,620 DEBUG HandlerThread:7281 [handler.py:handle_request():158] handle_request: defer +2024-05-14 10:30:57,620 INFO HandlerThread:7281 [handler.py:handle_request_defer():184] handle defer: 12 +2024-05-14 10:30:57,620 DEBUG SenderThread:7281 [sender.py:send_request():405] send_request: defer +2024-05-14 10:30:57,620 INFO SenderThread:7281 [sender.py:send_request_defer():609] handle sender defer: 12 +2024-05-14 10:30:57,620 INFO SenderThread:7281 [file_stream.py:finish():601] file stream finish called +2024-05-14 10:30:57,836 INFO SenderThread:7281 [file_stream.py:finish():605] file stream finish is done +2024-05-14 10:30:57,836 INFO SenderThread:7281 [sender.py:transition_state():613] send defer: 13 +2024-05-14 10:30:57,836 DEBUG HandlerThread:7281 [handler.py:handle_request():158] handle_request: defer +2024-05-14 10:30:57,836 INFO HandlerThread:7281 [handler.py:handle_request_defer():184] handle defer: 13 +2024-05-14 10:30:57,837 DEBUG SenderThread:7281 [sender.py:send_request():405] send_request: defer +2024-05-14 10:30:57,837 INFO SenderThread:7281 [sender.py:send_request_defer():609] handle sender defer: 13 +2024-05-14 10:30:57,837 INFO SenderThread:7281 [sender.py:transition_state():613] send defer: 14 +2024-05-14 10:30:57,837 DEBUG HandlerThread:7281 [handler.py:handle_request():158] handle_request: defer +2024-05-14 10:30:57,837 DEBUG SenderThread:7281 [sender.py:send():378] send: final +2024-05-14 10:30:57,837 INFO HandlerThread:7281 [handler.py:handle_request_defer():184] handle defer: 14 +2024-05-14 10:30:57,837 DEBUG SenderThread:7281 [sender.py:send():378] send: footer +2024-05-14 10:30:57,837 DEBUG SenderThread:7281 [sender.py:send_request():405] send_request: defer +2024-05-14 10:30:57,837 INFO SenderThread:7281 [sender.py:send_request_defer():609] handle sender defer: 14 +2024-05-14 10:30:57,838 DEBUG HandlerThread:7281 [handler.py:handle_request():158] handle_request: poll_exit +2024-05-14 10:30:57,838 DEBUG SenderThread:7281 [sender.py:send_request():405] send_request: poll_exit +2024-05-14 10:30:57,838 DEBUG HandlerThread:7281 [handler.py:handle_request():158] handle_request: poll_exit +2024-05-14 10:30:57,838 DEBUG SenderThread:7281 [sender.py:send_request():405] send_request: poll_exit +2024-05-14 10:30:57,839 DEBUG HandlerThread:7281 [handler.py:handle_request():158] handle_request: server_info +2024-05-14 10:30:57,839 DEBUG SenderThread:7281 [sender.py:send_request():405] send_request: server_info +2024-05-14 10:30:57,840 DEBUG HandlerThread:7281 [handler.py:handle_request():158] handle_request: get_summary +2024-05-14 10:30:57,840 DEBUG HandlerThread:7281 [handler.py:handle_request():158] handle_request: sampled_history +2024-05-14 10:30:57,840 DEBUG HandlerThread:7281 [handler.py:handle_request():158] handle_request: internal_messages +2024-05-14 10:30:57,901 INFO MainThread:7281 [wandb_run.py:_footer_history_summary_info():3994] rendering history +2024-05-14 10:30:57,902 INFO MainThread:7281 [wandb_run.py:_footer_history_summary_info():4026] rendering summary +2024-05-14 10:30:57,902 INFO MainThread:7281 [wandb_run.py:_footer_sync_info():3953] logging synced files +2024-05-14 10:30:57,902 DEBUG HandlerThread:7281 [handler.py:handle_request():158] handle_request: shutdown +2024-05-14 10:30:57,902 INFO HandlerThread:7281 [handler.py:finish():882] shutting down handler +2024-05-14 10:30:58,839 INFO WriterThread:7281 [datastore.py:close():296] close: /data/cronscript/lm-evaluation-harness/wandb/run-20240514_103049-lsh767q1/run-lsh767q1.wandb +2024-05-14 10:30:58,901 INFO SenderThread:7281 [sender.py:finish():1545] shutting down sender +2024-05-14 10:30:58,901 INFO SenderThread:7281 [file_pusher.py:finish():169] shutting down file pusher +2024-05-14 10:30:58,901 INFO SenderThread:7281 [file_pusher.py:join():175] waiting for file pusher diff --git a/lm-evaluation-harness/wandb/run-20240514_103049-lsh767q1/logs/debug.log b/lm-evaluation-harness/wandb/run-20240514_103049-lsh767q1/logs/debug.log new file mode 100644 index 0000000000000000000000000000000000000000..586cfcdc252f76bd21c6512c1a40d232e1dae739 --- /dev/null +++ b/lm-evaluation-harness/wandb/run-20240514_103049-lsh767q1/logs/debug.log @@ -0,0 +1,29 @@ +2024-05-14 10:30:49,698 INFO MainThread:6968 [wandb_setup.py:_flush():76] Current SDK version is 0.17.0 +2024-05-14 10:30:49,698 INFO MainThread:6968 [wandb_setup.py:_flush():76] Configure stats pid to 6968 +2024-05-14 10:30:49,698 INFO MainThread:6968 [wandb_setup.py:_flush():76] Loading settings from /root/.config/wandb/settings +2024-05-14 10:30:49,698 INFO MainThread:6968 [wandb_setup.py:_flush():76] Loading settings from /data/cronscript/lm-evaluation-harness/wandb/settings +2024-05-14 10:30:49,698 INFO MainThread:6968 [wandb_setup.py:_flush():76] Loading settings from environment variables: {} +2024-05-14 10:30:49,698 INFO MainThread:6968 [wandb_setup.py:_flush():76] Applying setup settings: {'_disable_service': False} +2024-05-14 10:30:49,698 WARNING MainThread:6968 [wandb_setup.py:_flush():76] Could not find program at -m lm_eval.__main__ +2024-05-14 10:30:49,698 INFO MainThread:6968 [wandb_setup.py:_flush():76] Inferring run settings from compute environment: {'program_relpath': None, 'program': '-m lm_eval.__main__'} +2024-05-14 10:30:49,698 INFO MainThread:6968 [wandb_setup.py:_flush():76] Applying login settings: {} +2024-05-14 10:30:49,698 INFO MainThread:6968 [wandb_init.py:_log_setup():520] Logging user logs to /data/cronscript/lm-evaluation-harness/wandb/run-20240514_103049-lsh767q1/logs/debug.log +2024-05-14 10:30:49,698 INFO MainThread:6968 [wandb_init.py:_log_setup():521] Logging internal logs to /data/cronscript/lm-evaluation-harness/wandb/run-20240514_103049-lsh767q1/logs/debug-internal.log +2024-05-14 10:30:49,698 INFO MainThread:6968 [wandb_init.py:init():560] calling init triggers +2024-05-14 10:30:49,698 INFO MainThread:6968 [wandb_init.py:init():567] wandb.init called with sweep_config: {} +config: {} +2024-05-14 10:30:49,698 INFO MainThread:6968 [wandb_init.py:init():610] starting backend +2024-05-14 10:30:49,698 INFO MainThread:6968 [wandb_init.py:init():614] setting up manager +2024-05-14 10:30:49,700 INFO MainThread:6968 [backend.py:_multiprocessing_setup():105] multiprocessing start_methods=fork,spawn,forkserver, using: spawn +2024-05-14 10:30:49,701 INFO MainThread:6968 [wandb_init.py:init():622] backend started and connected +2024-05-14 10:30:49,703 INFO MainThread:6968 [wandb_init.py:init():711] updated telemetry +2024-05-14 10:30:49,714 INFO MainThread:6968 [wandb_init.py:init():744] communicating run to backend with 90.0 second timeout +2024-05-14 10:30:49,930 INFO MainThread:6968 [wandb_run.py:_on_init():2396] communicating current version +2024-05-14 10:30:50,008 INFO MainThread:6968 [wandb_run.py:_on_init():2405] got version response +2024-05-14 10:30:50,008 INFO MainThread:6968 [wandb_init.py:init():795] starting run threads in backend +2024-05-14 10:30:50,208 INFO MainThread:6968 [wandb_run.py:_console_start():2374] atexit reg +2024-05-14 10:30:50,208 INFO MainThread:6968 [wandb_run.py:_redirect():2229] redirect: wrap_raw +2024-05-14 10:30:50,208 INFO MainThread:6968 [wandb_run.py:_redirect():2294] Wrapping output streams. +2024-05-14 10:30:50,208 INFO MainThread:6968 [wandb_run.py:_redirect():2319] Redirects installed. +2024-05-14 10:30:50,209 INFO MainThread:6968 [wandb_init.py:init():838] run started, returning control to user process +2024-05-14 10:30:58,903 WARNING MsgRouterThr:6968 [router.py:message_loop():77] message_loop has been closed diff --git a/lm-evaluation-harness/wandb/run-20240514_103049-lsh767q1/run-lsh767q1.wandb b/lm-evaluation-harness/wandb/run-20240514_103049-lsh767q1/run-lsh767q1.wandb new file mode 100644 index 0000000000000000000000000000000000000000..6fb78fa62caab81da326f5b8bedea3b54bd57cc0 Binary files /dev/null and b/lm-evaluation-harness/wandb/run-20240514_103049-lsh767q1/run-lsh767q1.wandb differ diff --git a/lm-evaluation-harness/wandb/run-20240514_103117-q880yywu/files/config.yaml b/lm-evaluation-harness/wandb/run-20240514_103117-q880yywu/files/config.yaml new file mode 100644 index 0000000000000000000000000000000000000000..57e40fdeb557f8826a74919741b8f08236b8c7ad --- /dev/null +++ b/lm-evaluation-harness/wandb/run-20240514_103117-q880yywu/files/config.yaml @@ -0,0 +1,43 @@ +wandb_version: 1 + +_wandb: + desc: null + value: + python_version: 3.10.12 + cli_version: 0.17.0 + framework: huggingface + huggingface_version: 4.40.2 + is_jupyter_run: false + is_kaggle_kernel: false + start_time: 1715682677 + t: + 1: + - 1 + - 5 + - 11 + - 49 + - 51 + - 53 + - 55 + - 71 + - 98 + - 100 + 2: + - 1 + - 5 + - 11 + - 49 + - 51 + - 53 + - 55 + - 71 + - 98 + - 100 + 3: + - 23 + 4: 3.10.12 + 5: 0.17.0 + 6: 4.40.2 + 8: + - 5 + 13: linux-x86_64 diff --git a/lm-evaluation-harness/wandb/run-20240514_103117-q880yywu/files/wandb-metadata.json b/lm-evaluation-harness/wandb/run-20240514_103117-q880yywu/files/wandb-metadata.json new file mode 100644 index 0000000000000000000000000000000000000000..40314f560c4fef0161155e0c88977fc3d08566a5 --- /dev/null +++ b/lm-evaluation-harness/wandb/run-20240514_103117-q880yywu/files/wandb-metadata.json @@ -0,0 +1,810 @@ +{ + "os": "Linux-5.15.0-92-generic-x86_64-with-glibc2.35", + "python": "3.10.12", + "heartbeatAt": "2024-05-14T10:31:17.487006", + "startedAt": "2024-05-14T10:31:17.103840", + "docker": null, + "cuda": null, + "args": [ + "--model", + "hf", + "--model_args", + "pretrained=/data/cronscript/ckpts//hf_ckpt//global_step20", + "--tasks", + "indiccopa-hi", + "--batch_size", + "auto", + "--wandb_args", + "project=bharatgpt" + ], + "state": "running", + "program": "-m lm_eval.__main__", + "codePathLocal": null, + "git": { + "remote": "https://github.com/EleutherAI/lm-evaluation-harness", + "commit": null + }, + "email": null, + "root": "/data/cronscript/lm-evaluation-harness", + "host": "vizzhy-150-3", + "username": "root", + "executable": "/usr/bin/python3", + "cpu_count": 76, + "cpu_count_logical": 152, + "cpu_freq": { + "current": 3394.130980263158, + "min": 800.0, + "max": 3400.0 + }, + "cpu_freq_per_core": [ + { + "current": 3300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3300.013, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3299.996, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3299.996, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + } + ], + "disk": { + "/": { + "total": 866.4415092468262, + "used": 76.92277145385742 + } + }, + "memory": { + "total": 1007.5000267028809 + } +} diff --git a/lm-evaluation-harness/wandb/run-20240514_103117-q880yywu/files/wandb-summary.json b/lm-evaluation-harness/wandb/run-20240514_103117-q880yywu/files/wandb-summary.json new file mode 100644 index 0000000000000000000000000000000000000000..e682bae6b5eaeba8295fd0fffdc51474a259249e --- /dev/null +++ b/lm-evaluation-harness/wandb/run-20240514_103117-q880yywu/files/wandb-summary.json @@ -0,0 +1 @@ +{"_wandb": {"runtime": 5}} \ No newline at end of file diff --git a/lm-evaluation-harness/wandb/run-20240514_103117-q880yywu/logs/debug-internal.log b/lm-evaluation-harness/wandb/run-20240514_103117-q880yywu/logs/debug-internal.log new file mode 100644 index 0000000000000000000000000000000000000000..6cedb92765b93171a9b6211553b5cd14b56d3d2e --- /dev/null +++ b/lm-evaluation-harness/wandb/run-20240514_103117-q880yywu/logs/debug-internal.log @@ -0,0 +1,181 @@ +2024-05-14 10:31:17,115 INFO StreamThr :9976 [internal.py:wandb_internal():85] W&B internal server running at pid: 9976, started at: 2024-05-14 10:31:17.114510 +2024-05-14 10:31:17,117 DEBUG HandlerThread:9976 [handler.py:handle_request():158] handle_request: status +2024-05-14 10:31:17,118 INFO WriterThread:9976 [datastore.py:open_for_write():87] open: /data/cronscript/lm-evaluation-harness/wandb/run-20240514_103117-q880yywu/run-q880yywu.wandb +2024-05-14 10:31:17,119 DEBUG SenderThread:9976 [sender.py:send():378] send: header +2024-05-14 10:31:17,127 DEBUG SenderThread:9976 [sender.py:send():378] send: run +2024-05-14 10:31:17,336 INFO SenderThread:9976 [dir_watcher.py:__init__():211] watching files in: /data/cronscript/lm-evaluation-harness/wandb/run-20240514_103117-q880yywu/files +2024-05-14 10:31:17,336 INFO SenderThread:9976 [sender.py:_start_run_threads():1123] run started: q880yywu with start time 1715682677.114269 +2024-05-14 10:31:17,352 DEBUG HandlerThread:9976 [handler.py:handle_request():158] handle_request: check_version +2024-05-14 10:31:17,352 DEBUG SenderThread:9976 [sender.py:send_request():405] send_request: check_version +2024-05-14 10:31:17,435 DEBUG HandlerThread:9976 [handler.py:handle_request():158] handle_request: run_start +2024-05-14 10:31:17,436 DEBUG HandlerThread:9976 [system_info.py:__init__():26] System info init +2024-05-14 10:31:17,436 DEBUG HandlerThread:9976 [system_info.py:__init__():41] System info init done +2024-05-14 10:31:17,436 INFO HandlerThread:9976 [system_monitor.py:start():194] Starting system monitor +2024-05-14 10:31:17,436 INFO SystemMonitor:9976 [system_monitor.py:_start():158] Starting system asset monitoring threads +2024-05-14 10:31:17,437 INFO HandlerThread:9976 [system_monitor.py:probe():214] Collecting system info +2024-05-14 10:31:17,437 INFO SystemMonitor:9976 [interfaces.py:start():188] Started cpu monitoring +2024-05-14 10:31:17,438 INFO SystemMonitor:9976 [interfaces.py:start():188] Started disk monitoring +2024-05-14 10:31:17,438 INFO SystemMonitor:9976 [interfaces.py:start():188] Started memory monitoring +2024-05-14 10:31:17,438 INFO SystemMonitor:9976 [interfaces.py:start():188] Started network monitoring +2024-05-14 10:31:17,486 DEBUG HandlerThread:9976 [system_info.py:probe():150] Probing system +2024-05-14 10:31:17,494 DEBUG HandlerThread:9976 [system_info.py:_probe_git():135] Probing git +2024-05-14 10:31:17,513 ERROR HandlerThread:9976 [gitlib.py:root():92] git root error: Cmd('git') failed due to: exit code(128) + cmdline: git rev-parse --show-toplevel + stderr: 'fatal: detected dubious ownership in repository at '/data/cronscript/lm-evaluation-harness' +To add an exception for this directory, call: + + git config --global --add safe.directory /data/cronscript/lm-evaluation-harness' +2024-05-14 10:31:17,513 DEBUG HandlerThread:9976 [system_info.py:_probe_git():143] Probing git done +2024-05-14 10:31:17,513 DEBUG HandlerThread:9976 [system_info.py:probe():198] Probing system done +2024-05-14 10:31:17,513 DEBUG HandlerThread:9976 [system_monitor.py:probe():223] {'os': 'Linux-5.15.0-92-generic-x86_64-with-glibc2.35', 'python': '3.10.12', 'heartbeatAt': '2024-05-14T10:31:17.487006', 'startedAt': '2024-05-14T10:31:17.103840', 'docker': None, 'cuda': None, 'args': ('--model', 'hf', '--model_args', 'pretrained=/data/cronscript/ckpts//hf_ckpt//global_step20', '--tasks', 'indiccopa-hi', '--batch_size', 'auto', '--wandb_args', 'project=bharatgpt'), 'state': 'running', 'program': '-m lm_eval.__main__', 'codePathLocal': None, 'git': {'remote': 'https://github.com/EleutherAI/lm-evaluation-harness', 'commit': None}, 'email': None, 'root': '/data/cronscript/lm-evaluation-harness', 'host': 'vizzhy-150-3', 'username': 'root', 'executable': '/usr/bin/python3', 'cpu_count': 76, 'cpu_count_logical': 152, 'cpu_freq': {'current': 3394.130980263158, 'min': 800.0, 'max': 3400.0}, 'cpu_freq_per_core': [{'current': 3300.0, 'min': 800.0, 'max': 3400.0}, {'current': 3300.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3300.013, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3300.0, 'min': 800.0, 'max': 3400.0}, {'current': 3299.996, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3300.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3299.996, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}], 'disk': {'/': {'total': 866.4415092468262, 'used': 76.92277145385742}}, 'memory': {'total': 1007.5000267028809}} +2024-05-14 10:31:17,514 INFO HandlerThread:9976 [system_monitor.py:probe():224] Finished collecting system info +2024-05-14 10:31:17,514 INFO HandlerThread:9976 [system_monitor.py:probe():227] Publishing system info +2024-05-14 10:31:17,515 INFO HandlerThread:9976 [system_monitor.py:probe():229] Finished publishing system info +2024-05-14 10:31:17,519 DEBUG SenderThread:9976 [sender.py:send():378] send: files +2024-05-14 10:31:17,519 INFO SenderThread:9976 [sender.py:_save_file():1389] saving file wandb-metadata.json with policy now +2024-05-14 10:31:17,650 DEBUG HandlerThread:9976 [handler.py:handle_request():158] handle_request: python_packages +2024-05-14 10:31:17,650 DEBUG SenderThread:9976 [sender.py:send_request():405] send_request: python_packages +2024-05-14 10:31:17,651 DEBUG HandlerThread:9976 [handler.py:handle_request():158] handle_request: stop_status +2024-05-14 10:31:17,651 DEBUG SenderThread:9976 [sender.py:send_request():405] send_request: stop_status +2024-05-14 10:31:17,777 DEBUG SenderThread:9976 [sender.py:send():378] send: telemetry +2024-05-14 10:31:18,022 INFO wandb-upload_0:9976 [upload_job.py:push():130] Uploaded file /tmp/tmpxwy_l4azwandb/5j276bu7-wandb-metadata.json +2024-05-14 10:31:18,337 INFO Thread-12 :9976 [dir_watcher.py:_on_file_created():271] file/dir created: /data/cronscript/lm-evaluation-harness/wandb/run-20240514_103117-q880yywu/files/wandb-metadata.json +2024-05-14 10:31:18,337 INFO Thread-12 :9976 [dir_watcher.py:_on_file_created():271] file/dir created: /data/cronscript/lm-evaluation-harness/wandb/run-20240514_103117-q880yywu/files/output.log +2024-05-14 10:31:18,337 INFO Thread-12 :9976 [dir_watcher.py:_on_file_created():271] file/dir created: /data/cronscript/lm-evaluation-harness/wandb/run-20240514_103117-q880yywu/files/requirements.txt +2024-05-14 10:31:20,338 INFO Thread-12 :9976 [dir_watcher.py:_on_file_modified():288] file/dir modified: /data/cronscript/lm-evaluation-harness/wandb/run-20240514_103117-q880yywu/files/output.log +2024-05-14 10:31:23,057 DEBUG HandlerThread:9976 [handler.py:handle_request():158] handle_request: status_report +2024-05-14 10:31:23,220 DEBUG SenderThread:9976 [sender.py:send():378] send: exit +2024-05-14 10:31:23,220 INFO SenderThread:9976 [sender.py:send_exit():585] handling exit code: 1 +2024-05-14 10:31:23,220 INFO SenderThread:9976 [sender.py:send_exit():587] handling runtime: 5 +2024-05-14 10:31:23,221 INFO SenderThread:9976 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end +2024-05-14 10:31:23,221 INFO SenderThread:9976 [sender.py:send_exit():593] send defer +2024-05-14 10:31:23,221 DEBUG HandlerThread:9976 [handler.py:handle_request():158] handle_request: defer +2024-05-14 10:31:23,221 INFO HandlerThread:9976 [handler.py:handle_request_defer():184] handle defer: 0 +2024-05-14 10:31:23,222 DEBUG SenderThread:9976 [sender.py:send_request():405] send_request: defer +2024-05-14 10:31:23,222 INFO SenderThread:9976 [sender.py:send_request_defer():609] handle sender defer: 0 +2024-05-14 10:31:23,222 INFO SenderThread:9976 [sender.py:transition_state():613] send defer: 1 +2024-05-14 10:31:23,222 DEBUG HandlerThread:9976 [handler.py:handle_request():158] handle_request: defer +2024-05-14 10:31:23,222 INFO HandlerThread:9976 [handler.py:handle_request_defer():184] handle defer: 1 +2024-05-14 10:31:23,222 DEBUG SenderThread:9976 [sender.py:send_request():405] send_request: defer +2024-05-14 10:31:23,222 INFO SenderThread:9976 [sender.py:send_request_defer():609] handle sender defer: 1 +2024-05-14 10:31:23,222 INFO SenderThread:9976 [sender.py:transition_state():613] send defer: 2 +2024-05-14 10:31:23,222 DEBUG HandlerThread:9976 [handler.py:handle_request():158] handle_request: defer +2024-05-14 10:31:23,222 INFO HandlerThread:9976 [handler.py:handle_request_defer():184] handle defer: 2 +2024-05-14 10:31:23,222 INFO HandlerThread:9976 [system_monitor.py:finish():203] Stopping system monitor +2024-05-14 10:31:23,222 DEBUG SystemMonitor:9976 [system_monitor.py:_start():172] Starting system metrics aggregation loop +2024-05-14 10:31:23,223 INFO HandlerThread:9976 [interfaces.py:finish():200] Joined cpu monitor +2024-05-14 10:31:23,223 DEBUG SystemMonitor:9976 [system_monitor.py:_start():179] Finished system metrics aggregation loop +2024-05-14 10:31:23,223 INFO HandlerThread:9976 [interfaces.py:finish():200] Joined disk monitor +2024-05-14 10:31:23,223 DEBUG SystemMonitor:9976 [system_monitor.py:_start():183] Publishing last batch of metrics +2024-05-14 10:31:23,223 INFO HandlerThread:9976 [interfaces.py:finish():200] Joined memory monitor +2024-05-14 10:31:23,224 INFO HandlerThread:9976 [interfaces.py:finish():200] Joined network monitor +2024-05-14 10:31:23,225 DEBUG SenderThread:9976 [sender.py:send_request():405] send_request: defer +2024-05-14 10:31:23,225 INFO SenderThread:9976 [sender.py:send_request_defer():609] handle sender defer: 2 +2024-05-14 10:31:23,225 INFO SenderThread:9976 [sender.py:transition_state():613] send defer: 3 +2024-05-14 10:31:23,225 DEBUG SenderThread:9976 [sender.py:send():378] send: stats +2024-05-14 10:31:23,225 DEBUG HandlerThread:9976 [handler.py:handle_request():158] handle_request: defer +2024-05-14 10:31:23,225 INFO HandlerThread:9976 [handler.py:handle_request_defer():184] handle defer: 3 +2024-05-14 10:31:23,226 DEBUG SenderThread:9976 [sender.py:send_request():405] send_request: defer +2024-05-14 10:31:23,226 INFO SenderThread:9976 [sender.py:send_request_defer():609] handle sender defer: 3 +2024-05-14 10:31:23,226 INFO SenderThread:9976 [sender.py:transition_state():613] send defer: 4 +2024-05-14 10:31:23,226 DEBUG HandlerThread:9976 [handler.py:handle_request():158] handle_request: defer +2024-05-14 10:31:23,226 INFO HandlerThread:9976 [handler.py:handle_request_defer():184] handle defer: 4 +2024-05-14 10:31:23,226 DEBUG SenderThread:9976 [sender.py:send_request():405] send_request: defer +2024-05-14 10:31:23,226 INFO SenderThread:9976 [sender.py:send_request_defer():609] handle sender defer: 4 +2024-05-14 10:31:23,226 INFO SenderThread:9976 [sender.py:transition_state():613] send defer: 5 +2024-05-14 10:31:23,226 DEBUG HandlerThread:9976 [handler.py:handle_request():158] handle_request: defer +2024-05-14 10:31:23,226 INFO HandlerThread:9976 [handler.py:handle_request_defer():184] handle defer: 5 +2024-05-14 10:31:23,226 DEBUG SenderThread:9976 [sender.py:send():378] send: summary +2024-05-14 10:31:23,227 INFO SenderThread:9976 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end +2024-05-14 10:31:23,227 DEBUG SenderThread:9976 [sender.py:send_request():405] send_request: defer +2024-05-14 10:31:23,227 INFO SenderThread:9976 [sender.py:send_request_defer():609] handle sender defer: 5 +2024-05-14 10:31:23,227 INFO SenderThread:9976 [sender.py:transition_state():613] send defer: 6 +2024-05-14 10:31:23,227 DEBUG HandlerThread:9976 [handler.py:handle_request():158] handle_request: defer +2024-05-14 10:31:23,227 INFO HandlerThread:9976 [handler.py:handle_request_defer():184] handle defer: 6 +2024-05-14 10:31:23,227 DEBUG SenderThread:9976 [sender.py:send_request():405] send_request: defer +2024-05-14 10:31:23,227 INFO SenderThread:9976 [sender.py:send_request_defer():609] handle sender defer: 6 +2024-05-14 10:31:23,230 DEBUG HandlerThread:9976 [handler.py:handle_request():158] handle_request: status_report +2024-05-14 10:31:23,340 INFO Thread-12 :9976 [dir_watcher.py:_on_file_created():271] file/dir created: /data/cronscript/lm-evaluation-harness/wandb/run-20240514_103117-q880yywu/files/wandb-summary.json +2024-05-14 10:31:23,577 INFO SenderThread:9976 [sender.py:transition_state():613] send defer: 7 +2024-05-14 10:31:23,577 DEBUG HandlerThread:9976 [handler.py:handle_request():158] handle_request: defer +2024-05-14 10:31:23,577 INFO HandlerThread:9976 [handler.py:handle_request_defer():184] handle defer: 7 +2024-05-14 10:31:23,577 DEBUG SenderThread:9976 [sender.py:send_request():405] send_request: defer +2024-05-14 10:31:23,577 INFO SenderThread:9976 [sender.py:send_request_defer():609] handle sender defer: 7 +2024-05-14 10:31:23,795 INFO SenderThread:9976 [sender.py:transition_state():613] send defer: 8 +2024-05-14 10:31:23,795 DEBUG HandlerThread:9976 [handler.py:handle_request():158] handle_request: defer +2024-05-14 10:31:23,796 INFO HandlerThread:9976 [handler.py:handle_request_defer():184] handle defer: 8 +2024-05-14 10:31:23,796 DEBUG SenderThread:9976 [sender.py:send_request():405] send_request: defer +2024-05-14 10:31:23,796 INFO SenderThread:9976 [sender.py:send_request_defer():609] handle sender defer: 8 +2024-05-14 10:31:23,796 INFO SenderThread:9976 [job_builder.py:build():432] Attempting to build job artifact +2024-05-14 10:31:23,796 INFO SenderThread:9976 [job_builder.py:_get_source_type():576] no source found +2024-05-14 10:31:23,796 INFO SenderThread:9976 [sender.py:transition_state():613] send defer: 9 +2024-05-14 10:31:23,796 DEBUG HandlerThread:9976 [handler.py:handle_request():158] handle_request: defer +2024-05-14 10:31:23,796 INFO HandlerThread:9976 [handler.py:handle_request_defer():184] handle defer: 9 +2024-05-14 10:31:23,796 DEBUG SenderThread:9976 [sender.py:send_request():405] send_request: defer +2024-05-14 10:31:23,796 INFO SenderThread:9976 [sender.py:send_request_defer():609] handle sender defer: 9 +2024-05-14 10:31:23,796 INFO SenderThread:9976 [dir_watcher.py:finish():358] shutting down directory watcher +2024-05-14 10:31:24,220 DEBUG HandlerThread:9976 [handler.py:handle_request():158] handle_request: poll_exit +2024-05-14 10:31:24,341 INFO SenderThread:9976 [dir_watcher.py:_on_file_modified():288] file/dir modified: /data/cronscript/lm-evaluation-harness/wandb/run-20240514_103117-q880yywu/files/config.yaml +2024-05-14 10:31:24,341 INFO SenderThread:9976 [dir_watcher.py:_on_file_modified():288] file/dir modified: /data/cronscript/lm-evaluation-harness/wandb/run-20240514_103117-q880yywu/files/output.log +2024-05-14 10:31:24,341 INFO SenderThread:9976 [dir_watcher.py:finish():388] scan: /data/cronscript/lm-evaluation-harness/wandb/run-20240514_103117-q880yywu/files +2024-05-14 10:31:24,341 INFO SenderThread:9976 [dir_watcher.py:finish():402] scan save: /data/cronscript/lm-evaluation-harness/wandb/run-20240514_103117-q880yywu/files/requirements.txt requirements.txt +2024-05-14 10:31:24,341 INFO SenderThread:9976 [dir_watcher.py:finish():402] scan save: /data/cronscript/lm-evaluation-harness/wandb/run-20240514_103117-q880yywu/files/wandb-metadata.json wandb-metadata.json +2024-05-14 10:31:24,341 INFO SenderThread:9976 [dir_watcher.py:finish():402] scan save: /data/cronscript/lm-evaluation-harness/wandb/run-20240514_103117-q880yywu/files/config.yaml config.yaml +2024-05-14 10:31:24,341 INFO SenderThread:9976 [dir_watcher.py:finish():402] scan save: /data/cronscript/lm-evaluation-harness/wandb/run-20240514_103117-q880yywu/files/output.log output.log +2024-05-14 10:31:24,341 INFO SenderThread:9976 [dir_watcher.py:finish():402] scan save: /data/cronscript/lm-evaluation-harness/wandb/run-20240514_103117-q880yywu/files/wandb-summary.json wandb-summary.json +2024-05-14 10:31:24,343 INFO SenderThread:9976 [sender.py:transition_state():613] send defer: 10 +2024-05-14 10:31:24,344 DEBUG SenderThread:9976 [sender.py:send_request():405] send_request: poll_exit +2024-05-14 10:31:24,345 DEBUG HandlerThread:9976 [handler.py:handle_request():158] handle_request: defer +2024-05-14 10:31:24,346 INFO HandlerThread:9976 [handler.py:handle_request_defer():184] handle defer: 10 +2024-05-14 10:31:24,348 DEBUG SenderThread:9976 [sender.py:send_request():405] send_request: defer +2024-05-14 10:31:24,348 INFO SenderThread:9976 [sender.py:send_request_defer():609] handle sender defer: 10 +2024-05-14 10:31:24,348 INFO SenderThread:9976 [file_pusher.py:finish():169] shutting down file pusher +2024-05-14 10:31:24,570 INFO wandb-upload_1:9976 [upload_job.py:push():130] Uploaded file /data/cronscript/lm-evaluation-harness/wandb/run-20240514_103117-q880yywu/files/requirements.txt +2024-05-14 10:31:24,857 INFO wandb-upload_0:9976 [upload_job.py:push():130] Uploaded file /data/cronscript/lm-evaluation-harness/wandb/run-20240514_103117-q880yywu/files/config.yaml +2024-05-14 10:31:24,874 INFO wandb-upload_3:9976 [upload_job.py:push():130] Uploaded file /data/cronscript/lm-evaluation-harness/wandb/run-20240514_103117-q880yywu/files/wandb-summary.json +2024-05-14 10:31:24,903 INFO wandb-upload_2:9976 [upload_job.py:push():130] Uploaded file /data/cronscript/lm-evaluation-harness/wandb/run-20240514_103117-q880yywu/files/output.log +2024-05-14 10:31:25,104 INFO Thread-11 (_thread_body):9976 [sender.py:transition_state():613] send defer: 11 +2024-05-14 10:31:25,104 DEBUG HandlerThread:9976 [handler.py:handle_request():158] handle_request: defer +2024-05-14 10:31:25,104 INFO HandlerThread:9976 [handler.py:handle_request_defer():184] handle defer: 11 +2024-05-14 10:31:25,105 DEBUG SenderThread:9976 [sender.py:send_request():405] send_request: defer +2024-05-14 10:31:25,105 INFO SenderThread:9976 [sender.py:send_request_defer():609] handle sender defer: 11 +2024-05-14 10:31:25,105 INFO SenderThread:9976 [file_pusher.py:join():175] waiting for file pusher +2024-05-14 10:31:25,105 INFO SenderThread:9976 [sender.py:transition_state():613] send defer: 12 +2024-05-14 10:31:25,105 DEBUG HandlerThread:9976 [handler.py:handle_request():158] handle_request: defer +2024-05-14 10:31:25,105 INFO HandlerThread:9976 [handler.py:handle_request_defer():184] handle defer: 12 +2024-05-14 10:31:25,105 DEBUG SenderThread:9976 [sender.py:send_request():405] send_request: defer +2024-05-14 10:31:25,105 INFO SenderThread:9976 [sender.py:send_request_defer():609] handle sender defer: 12 +2024-05-14 10:31:25,105 INFO SenderThread:9976 [file_stream.py:finish():601] file stream finish called +2024-05-14 10:31:25,220 DEBUG HandlerThread:9976 [handler.py:handle_request():158] handle_request: poll_exit +2024-05-14 10:31:25,587 INFO SenderThread:9976 [file_stream.py:finish():605] file stream finish is done +2024-05-14 10:31:25,588 INFO SenderThread:9976 [sender.py:transition_state():613] send defer: 13 +2024-05-14 10:31:25,588 DEBUG SenderThread:9976 [sender.py:send_request():405] send_request: poll_exit +2024-05-14 10:31:25,588 DEBUG HandlerThread:9976 [handler.py:handle_request():158] handle_request: defer +2024-05-14 10:31:25,588 INFO HandlerThread:9976 [handler.py:handle_request_defer():184] handle defer: 13 +2024-05-14 10:31:25,588 DEBUG SenderThread:9976 [sender.py:send_request():405] send_request: defer +2024-05-14 10:31:25,588 INFO SenderThread:9976 [sender.py:send_request_defer():609] handle sender defer: 13 +2024-05-14 10:31:25,588 INFO SenderThread:9976 [sender.py:transition_state():613] send defer: 14 +2024-05-14 10:31:25,588 DEBUG HandlerThread:9976 [handler.py:handle_request():158] handle_request: defer +2024-05-14 10:31:25,588 INFO HandlerThread:9976 [handler.py:handle_request_defer():184] handle defer: 14 +2024-05-14 10:31:25,589 DEBUG SenderThread:9976 [sender.py:send():378] send: final +2024-05-14 10:31:25,589 DEBUG SenderThread:9976 [sender.py:send():378] send: footer +2024-05-14 10:31:25,589 DEBUG SenderThread:9976 [sender.py:send_request():405] send_request: defer +2024-05-14 10:31:25,589 INFO SenderThread:9976 [sender.py:send_request_defer():609] handle sender defer: 14 +2024-05-14 10:31:25,589 DEBUG HandlerThread:9976 [handler.py:handle_request():158] handle_request: poll_exit +2024-05-14 10:31:25,589 DEBUG SenderThread:9976 [sender.py:send_request():405] send_request: poll_exit +2024-05-14 10:31:25,589 DEBUG HandlerThread:9976 [handler.py:handle_request():158] handle_request: poll_exit +2024-05-14 10:31:25,590 DEBUG HandlerThread:9976 [handler.py:handle_request():158] handle_request: server_info +2024-05-14 10:31:25,590 DEBUG SenderThread:9976 [sender.py:send_request():405] send_request: poll_exit +2024-05-14 10:31:25,590 DEBUG HandlerThread:9976 [handler.py:handle_request():158] handle_request: get_summary +2024-05-14 10:31:25,590 DEBUG SenderThread:9976 [sender.py:send_request():405] send_request: server_info +2024-05-14 10:31:25,590 DEBUG HandlerThread:9976 [handler.py:handle_request():158] handle_request: sampled_history +2024-05-14 10:31:25,591 DEBUG HandlerThread:9976 [handler.py:handle_request():158] handle_request: internal_messages +2024-05-14 10:31:25,655 INFO MainThread:9976 [wandb_run.py:_footer_history_summary_info():3994] rendering history +2024-05-14 10:31:25,655 INFO MainThread:9976 [wandb_run.py:_footer_history_summary_info():4026] rendering summary +2024-05-14 10:31:25,655 INFO MainThread:9976 [wandb_run.py:_footer_sync_info():3953] logging synced files +2024-05-14 10:31:25,655 DEBUG HandlerThread:9976 [handler.py:handle_request():158] handle_request: shutdown +2024-05-14 10:31:25,655 INFO HandlerThread:9976 [handler.py:finish():882] shutting down handler +2024-05-14 10:31:26,590 INFO WriterThread:9976 [datastore.py:close():296] close: /data/cronscript/lm-evaluation-harness/wandb/run-20240514_103117-q880yywu/run-q880yywu.wandb +2024-05-14 10:31:26,655 INFO SenderThread:9976 [sender.py:finish():1545] shutting down sender +2024-05-14 10:31:26,655 INFO SenderThread:9976 [file_pusher.py:finish():169] shutting down file pusher +2024-05-14 10:31:26,655 INFO SenderThread:9976 [file_pusher.py:join():175] waiting for file pusher diff --git a/lm-evaluation-harness/wandb/run-20240514_103117-q880yywu/logs/debug.log b/lm-evaluation-harness/wandb/run-20240514_103117-q880yywu/logs/debug.log new file mode 100644 index 0000000000000000000000000000000000000000..31d70d0f9d8c27afcc1d856846a8ea7a9634c5d1 --- /dev/null +++ b/lm-evaluation-harness/wandb/run-20240514_103117-q880yywu/logs/debug.log @@ -0,0 +1,28 @@ +2024-05-14 10:31:17,111 INFO MainThread:8780 [wandb_setup.py:_flush():76] Current SDK version is 0.17.0 +2024-05-14 10:31:17,111 INFO MainThread:8780 [wandb_setup.py:_flush():76] Configure stats pid to 8780 +2024-05-14 10:31:17,111 INFO MainThread:8780 [wandb_setup.py:_flush():76] Loading settings from /root/.config/wandb/settings +2024-05-14 10:31:17,111 INFO MainThread:8780 [wandb_setup.py:_flush():76] Loading settings from /data/cronscript/lm-evaluation-harness/wandb/settings +2024-05-14 10:31:17,111 INFO MainThread:8780 [wandb_setup.py:_flush():76] Loading settings from environment variables: {} +2024-05-14 10:31:17,111 INFO MainThread:8780 [wandb_setup.py:_flush():76] Applying setup settings: {'_disable_service': False} +2024-05-14 10:31:17,111 WARNING MainThread:8780 [wandb_setup.py:_flush():76] Could not find program at -m lm_eval.__main__ +2024-05-14 10:31:17,111 INFO MainThread:8780 [wandb_setup.py:_flush():76] Inferring run settings from compute environment: {'program_relpath': None, 'program': '-m lm_eval.__main__'} +2024-05-14 10:31:17,111 INFO MainThread:8780 [wandb_setup.py:_flush():76] Applying login settings: {} +2024-05-14 10:31:17,111 INFO MainThread:8780 [wandb_init.py:_log_setup():520] Logging user logs to /data/cronscript/lm-evaluation-harness/wandb/run-20240514_103117-q880yywu/logs/debug.log +2024-05-14 10:31:17,111 INFO MainThread:8780 [wandb_init.py:_log_setup():521] Logging internal logs to /data/cronscript/lm-evaluation-harness/wandb/run-20240514_103117-q880yywu/logs/debug-internal.log +2024-05-14 10:31:17,111 INFO MainThread:8780 [wandb_init.py:init():560] calling init triggers +2024-05-14 10:31:17,111 INFO MainThread:8780 [wandb_init.py:init():567] wandb.init called with sweep_config: {} +config: {} +2024-05-14 10:31:17,111 INFO MainThread:8780 [wandb_init.py:init():610] starting backend +2024-05-14 10:31:17,111 INFO MainThread:8780 [wandb_init.py:init():614] setting up manager +2024-05-14 10:31:17,113 INFO MainThread:8780 [backend.py:_multiprocessing_setup():105] multiprocessing start_methods=fork,spawn,forkserver, using: spawn +2024-05-14 10:31:17,114 INFO MainThread:8780 [wandb_init.py:init():622] backend started and connected +2024-05-14 10:31:17,116 INFO MainThread:8780 [wandb_init.py:init():711] updated telemetry +2024-05-14 10:31:17,127 INFO MainThread:8780 [wandb_init.py:init():744] communicating run to backend with 90.0 second timeout +2024-05-14 10:31:17,351 INFO MainThread:8780 [wandb_run.py:_on_init():2396] communicating current version +2024-05-14 10:31:17,430 INFO MainThread:8780 [wandb_run.py:_on_init():2405] got version response +2024-05-14 10:31:17,430 INFO MainThread:8780 [wandb_init.py:init():795] starting run threads in backend +2024-05-14 10:31:17,650 INFO MainThread:8780 [wandb_run.py:_console_start():2374] atexit reg +2024-05-14 10:31:17,650 INFO MainThread:8780 [wandb_run.py:_redirect():2229] redirect: wrap_raw +2024-05-14 10:31:17,651 INFO MainThread:8780 [wandb_run.py:_redirect():2294] Wrapping output streams. +2024-05-14 10:31:17,651 INFO MainThread:8780 [wandb_run.py:_redirect():2319] Redirects installed. +2024-05-14 10:31:17,652 INFO MainThread:8780 [wandb_init.py:init():838] run started, returning control to user process diff --git a/lm-evaluation-harness/wandb/run-20240514_163424-53eroz7t/files/config.yaml b/lm-evaluation-harness/wandb/run-20240514_163424-53eroz7t/files/config.yaml new file mode 100644 index 0000000000000000000000000000000000000000..f7dabc891edb6d0c65a5c4671e8f6d6510d0fbed --- /dev/null +++ b/lm-evaluation-harness/wandb/run-20240514_163424-53eroz7t/files/config.yaml @@ -0,0 +1,43 @@ +wandb_version: 1 + +_wandb: + desc: null + value: + python_version: 3.10.12 + cli_version: 0.17.0 + framework: huggingface + huggingface_version: 4.40.2 + is_jupyter_run: false + is_kaggle_kernel: false + start_time: 1715704464 + t: + 1: + - 1 + - 5 + - 11 + - 49 + - 51 + - 53 + - 55 + - 71 + - 98 + - 100 + 2: + - 1 + - 5 + - 11 + - 49 + - 51 + - 53 + - 55 + - 71 + - 98 + - 100 + 3: + - 23 + 4: 3.10.12 + 5: 0.17.0 + 6: 4.40.2 + 8: + - 5 + 13: linux-x86_64 diff --git a/lm-evaluation-harness/wandb/run-20240514_163424-53eroz7t/files/output.log b/lm-evaluation-harness/wandb/run-20240514_163424-53eroz7t/files/output.log new file mode 100644 index 0000000000000000000000000000000000000000..cb2087b883df3fab8f6fe004765344b0db5d1724 --- /dev/null +++ b/lm-evaluation-harness/wandb/run-20240514_163424-53eroz7t/files/output.log @@ -0,0 +1,28 @@ + +2024-05-14:16:34:25,071 INFO [__main__.py:251] Verbosity set to INFO +2024-05-14:16:34:30,618 INFO [__main__.py:335] Selected Tasks: ['indiccopa-hi'] +2024-05-14:16:34:30,622 INFO [evaluator.py:131] Setting random seed to 0 | Setting numpy seed to 1234 | Setting torch manual seed to 1234 +2024-05-14:16:34:30,622 INFO [evaluator.py:177] Initializing hf model, with arguments: {'pretrained': '/data/cronscript/ckpts//hf_ckpt//global_step100'} +/usr/local/lib/python3.10/dist-packages/habana_frameworks/torch/gpu_migration/core/register.py:145: UserWarning: "hpu:X" notation is not supported by Gaudi PyTorch intergration bridge. Please change to "hpu" without index (Triggered internally at /npu-stack/pytorch-integration/pytorch_helpers/lazy_to_backend.cpp:53.) + return func(*args, **kwargs) +/usr/local/lib/python3.10/dist-packages/habana_frameworks/torch/gpu_migration/torch/cuda/memory.py:36: UserWarning: No need to call empty_cache on HPU. It manages the memory internally in an effcient way. + warnings.warn( +[2024-05-14 16:34:41,217] [INFO] [real_accelerator.py:178:get_accelerator] Setting ds_accelerator to hpu (auto detect) +/usr/local/lib/python3.10/dist-packages/habana_frameworks/torch/hpu/__init__.py:158: UserWarning: torch.hpu.setDeterministic is deprecated and will be removed in next release. Please use torch.use_deterministic_algorithms instead. + warnings.warn( +You are using the default legacy behaviour of the . This is expected, and simply means that the `legacy` (previous) behavior will be used so nothing changes for you. If you want to use the new behaviour, set `legacy=False`. This should only be set if you understand what it means, and thoroughly read the reason why this was added as explained in https://github.com/huggingface/transformers/pull/24565 +2024-05-14:16:34:41,632 WARNING [task.py:763] [Task: indiccopa-hi] metric acc is defined, but aggregation is not. using default aggregation=mean +2024-05-14:16:34:41,632 WARNING [task.py:775] [Task: indiccopa-hi] metric acc is defined, but higher_is_better is not. using default higher_is_better=True +/usr/local/lib/python3.10/dist-packages/datasets/load.py:1486: FutureWarning: The repository for ai4bharat/IndicCOPA contains custom code which must be executed to correctly load the dataset. You can inspect the repository content at https://hf.co/datasets/ai4bharat/IndicCOPA +You can avoid this message in future by passing the argument `trust_remote_code=True`. +Passing `trust_remote_code=True` will be mandatory to load this dataset from the next major release of `datasets`. + warnings.warn( +Passed argument batch_size = auto:1. Detecting largest batch size +2024-05-14:16:34:42,903 WARNING [task.py:322] [Task: indiccopa-hi] has_training_docs and has_validation_docs are False, using test_docs as fewshot_docs but this is not recommended. +2024-05-14:16:34:42,904 WARNING [task.py:322] [Task: indiccopa-hi] has_training_docs and has_validation_docs are False, using test_docs as fewshot_docs but this is not recommended. +2024-05-14:16:34:42,928 INFO [task.py:395] Building contexts for indiccopa-hi on rank 1... +100%|█████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████| 56/56 [00:00<00:00, 88936.40it/s] +2024-05-14:16:34:44,718 INFO [evaluator.py:379] Running loglikelihood requests +Determined largest batch size: 64 +/usr/local/lib/python3.10/dist-packages/habana_frameworks/torch/gpu_migration/torch/cuda/memory.py:36: UserWarning: No need to call empty_cache on HPU. It manages the memory internally in an effcient way. + warnings.warn( \ No newline at end of file diff --git a/lm-evaluation-harness/wandb/run-20240514_163424-53eroz7t/files/requirements.txt b/lm-evaluation-harness/wandb/run-20240514_163424-53eroz7t/files/requirements.txt new file mode 100644 index 0000000000000000000000000000000000000000..d63edd781bd5bdbb7f67523ac1ba9f0f1ed392dc --- /dev/null +++ b/lm-evaluation-harness/wandb/run-20240514_163424-53eroz7t/files/requirements.txt @@ -0,0 +1,163 @@ +DataProperty==1.0.1 +GitPython==3.1.43 +Jinja2==3.1.3 +Markdown==3.6 +MarkupSafe==2.1.5 +Pillow-SIMD==7.0.0.post3 +PyYAML==6.0 +Werkzeug==3.0.2 +absl-py==2.1.0 +accelerate==0.30.1 +aiohttp==3.9.4 +aiosignal==1.3.1 +antlr4-python3-runtime==4.9.3 +anyio==4.3.0 +async-timeout==4.0.3 +attrs==23.2.0 +av==9.2.0 +cachetools==5.3.3 +certifi==2024.2.2 +cffi==1.15.1 +cfgv==3.4.0 +chardet==5.2.0 +charset-normalizer==3.3.2 +click==8.1.7 +cmake==3.29.2 +colorama==0.4.6 +datasets==2.19.1 +deepspeed==0.12.4+hpu.synapse.v1.15.1 +dill==0.3.8 +distlib==0.3.8 +distro==1.9.0 +docker-pycreds==0.4.0 +einops==0.8.0 +evaluate==0.4.2 +exceptiongroup==1.2.0 +expecttest==0.2.1 +filelock==3.13.4 +frozenlist==1.4.1 +fsspec==2024.3.1 +gitdb==4.0.11 +google-auth-oauthlib==0.4.6 +google-auth==2.29.0 +grpcio==1.62.1 +h11==0.14.0 +habana-media-loader==1.15.1.15 +habana-pyhlml==1.15.1.15 +habana-torch-dataloader==1.15.1.15 +habana-torch-plugin==1.15.1.15 +habana_gpu_migration==1.15.1.15 +habana_quantization_toolkit==1.15.1.15 +hjson==3.1.0 +httpcore==1.0.5 +httpx==0.27.0 +huggingface-hub==0.23.0 +identify==2.5.35 +idna==3.7 +importlib_resources==6.4.0 +iniconfig==2.0.0 +joblib==1.4.2 +jsonlines==4.0.0 +lightning-habana==1.4.0 +lightning-utilities==0.11.2 +lightning==2.2.0.post0 +lm_eval==0.3.0 +lm_eval==0.4.2 +lm_eval==0.4.2 +lm_eval==0.4.2 +mbstrdecoder==1.1.3 +more-itertools==10.2.0 +mpi4py==3.1.4 +mpmath==1.3.0 +multidict==6.0.5 +multiprocess==0.70.16 +networkx==3.3 +ninja==1.11.1.1 +nltk==3.8.1 +nodeenv==1.8.0 +numexpr==2.10.0 +numpy==1.23.5 +oauthlib==3.2.2 +omegaconf==2.3.0 +openai==1.29.0 +packaging==24.0 +pandas==2.0.1 +pathspec==0.12.1 +pathvalidate==3.2.0 +peft==0.10.0 +perfetto==0.7.0 +pip==22.0.2 +pip==23.3.1 +platformdirs==4.2.0 +pluggy==1.4.0 +portalocker==2.8.2 +pre-commit==3.3.3 +protobuf==3.20.3 +psutil==5.9.8 +py-cpuinfo==9.0.0 +pyarrow-hotfix==0.6 +pyarrow==16.0.0 +pyasn1==0.6.0 +pyasn1_modules==0.4.0 +pybind11==2.10.4 +pycountry==23.12.11 +pycparser==2.22 +pydantic==1.10.13 +pynvml==8.0.4 +pytablewriter==1.2.0 +pytest==8.1.1 +python-dateutil==2.9.0.post0 +pytorch-lightning==2.2.2 +pytz==2024.1 +regex==2023.5.5 +requests-oauthlib==2.0.0 +requests==2.31.0 +rouge_score==0.1.2 +rsa==4.9 +sacrebleu==1.5.0 +safetensors==0.4.3 +scikit-learn==1.4.2 +scipy==1.13.0 +sentencepiece==0.2.0 +sentry-sdk==2.1.1 +setproctitle==1.3.3 +setuptools==59.6.0 +setuptools==69.5.1 +six==1.16.0 +smmap==5.0.1 +sniffio==1.3.1 +sqlitedict==2.1.0 +symengine==0.11.0 +sympy==1.12 +tabledata==1.3.3 +tcolorpy==0.1.6 +tdqm==0.0.1 +tensorboard-data-server==0.6.1 +tensorboard-plugin-wit==1.8.1 +tensorboard==2.11.2 +threadpoolctl==3.5.0 +tokenizers==0.19.1 +tomli==2.0.1 +torch==2.2.0a0+git8964477 +torch_tb_profiler==0.4.0 +torchaudio==2.2.0+08901ad +torchdata==0.7.1+5e6f7b7 +torchmetrics==1.3.2 +torchtext==0.17.0+400da5c +torchvision==0.17.0+b2383d4 +tqdm-multiprocess==0.0.11 +tqdm==4.66.2 +transformers==4.40.2 +typepy==1.3.2 +typing_extensions==4.11.0 +tzdata==2024.1 +urllib3==1.26.18 +virtualenv==20.25.1 +wandb==0.17.0 +wheel==0.37.1 +wheel==0.43.0 +word2number==1.1 +xxhash==3.4.1 +yamllint==1.35.1 +yarl==1.9.4 +zstandard==0.22.0 \ No newline at end of file diff --git a/lm-evaluation-harness/wandb/run-20240514_163424-53eroz7t/files/wandb-metadata.json b/lm-evaluation-harness/wandb/run-20240514_163424-53eroz7t/files/wandb-metadata.json new file mode 100644 index 0000000000000000000000000000000000000000..ceb5165bc245c62e89319a9300063207b39acd7f --- /dev/null +++ b/lm-evaluation-harness/wandb/run-20240514_163424-53eroz7t/files/wandb-metadata.json @@ -0,0 +1,810 @@ +{ + "os": "Linux-5.15.0-92-generic-x86_64-with-glibc2.35", + "python": "3.10.12", + "heartbeatAt": "2024-05-14T16:34:24.936081", + "startedAt": "2024-05-14T16:34:24.428958", + "docker": null, + "cuda": null, + "args": [ + "--model", + "hf", + "--model_args", + "pretrained=/data/cronscript/ckpts//hf_ckpt//global_step100", + "--tasks", + "indiccopa-hi", + "--batch_size", + "auto", + "--wandb_args", + "project=bharatgpt,group=trial_expt" + ], + "state": "running", + "program": "-m lm_eval.__main__", + "codePathLocal": null, + "git": { + "remote": "https://github.com/EleutherAI/lm-evaluation-harness", + "commit": null + }, + "email": null, + "root": "/data/cronscript/lm-evaluation-harness", + "host": "vizzhy-150-3", + "username": "root", + "executable": "/usr/bin/python3", + "cpu_count": 76, + "cpu_count_logical": 152, + "cpu_freq": { + "current": 3394.7368157894734, + "min": 800.0, + "max": 3400.0 + }, + "cpu_freq_per_core": [ + { + "current": 3300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3300.002, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3313.888, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3210.166, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + } + ], + "disk": { + "/": { + "total": 866.4415092468262, + "used": 863.4841461181641 + } + }, + "memory": { + "total": 1007.5000267028809 + } +} diff --git a/lm-evaluation-harness/wandb/run-20240514_163424-53eroz7t/files/wandb-summary.json b/lm-evaluation-harness/wandb/run-20240514_163424-53eroz7t/files/wandb-summary.json new file mode 100644 index 0000000000000000000000000000000000000000..f54bb3d71ddd5d7fb0c66b39e940c1d03cfc7cd1 --- /dev/null +++ b/lm-evaluation-harness/wandb/run-20240514_163424-53eroz7t/files/wandb-summary.json @@ -0,0 +1 @@ +{"_wandb": {"runtime": 27}} \ No newline at end of file diff --git a/lm-evaluation-harness/wandb/run-20240514_163424-53eroz7t/logs/debug-internal.log b/lm-evaluation-harness/wandb/run-20240514_163424-53eroz7t/logs/debug-internal.log new file mode 100644 index 0000000000000000000000000000000000000000..e6fa8058c6acabe313c536c15e4834d39db78ad4 --- /dev/null +++ b/lm-evaluation-harness/wandb/run-20240514_163424-53eroz7t/logs/debug-internal.log @@ -0,0 +1,194 @@ +2024-05-14 16:34:24,440 INFO StreamThr :117696 [internal.py:wandb_internal():85] W&B internal server running at pid: 117696, started at: 2024-05-14 16:34:24.440004 +2024-05-14 16:34:24,443 DEBUG HandlerThread:117696 [handler.py:handle_request():158] handle_request: status +2024-05-14 16:34:24,444 INFO WriterThread:117696 [datastore.py:open_for_write():87] open: /data/cronscript/lm-evaluation-harness/wandb/run-20240514_163424-53eroz7t/run-53eroz7t.wandb +2024-05-14 16:34:24,445 DEBUG SenderThread:117696 [sender.py:send():378] send: header +2024-05-14 16:34:24,453 DEBUG SenderThread:117696 [sender.py:send():378] send: run +2024-05-14 16:34:24,754 INFO SenderThread:117696 [dir_watcher.py:__init__():211] watching files in: /data/cronscript/lm-evaluation-harness/wandb/run-20240514_163424-53eroz7t/files +2024-05-14 16:34:24,754 INFO SenderThread:117696 [sender.py:_start_run_threads():1123] run started: 53eroz7t with start time 1715704464.439408 +2024-05-14 16:34:24,758 DEBUG HandlerThread:117696 [handler.py:handle_request():158] handle_request: check_version +2024-05-14 16:34:24,759 DEBUG SenderThread:117696 [sender.py:send_request():405] send_request: check_version +2024-05-14 16:34:24,847 DEBUG HandlerThread:117696 [handler.py:handle_request():158] handle_request: run_start +2024-05-14 16:34:24,849 DEBUG HandlerThread:117696 [system_info.py:__init__():26] System info init +2024-05-14 16:34:24,849 DEBUG HandlerThread:117696 [system_info.py:__init__():41] System info init done +2024-05-14 16:34:24,849 INFO HandlerThread:117696 [system_monitor.py:start():194] Starting system monitor +2024-05-14 16:34:24,849 INFO SystemMonitor:117696 [system_monitor.py:_start():158] Starting system asset monitoring threads +2024-05-14 16:34:24,849 INFO HandlerThread:117696 [system_monitor.py:probe():214] Collecting system info +2024-05-14 16:34:24,850 INFO SystemMonitor:117696 [interfaces.py:start():188] Started cpu monitoring +2024-05-14 16:34:24,852 INFO SystemMonitor:117696 [interfaces.py:start():188] Started disk monitoring +2024-05-14 16:34:24,853 INFO SystemMonitor:117696 [interfaces.py:start():188] Started memory monitoring +2024-05-14 16:34:24,854 INFO SystemMonitor:117696 [interfaces.py:start():188] Started network monitoring +2024-05-14 16:34:24,936 DEBUG HandlerThread:117696 [system_info.py:probe():150] Probing system +2024-05-14 16:34:24,944 DEBUG HandlerThread:117696 [system_info.py:_probe_git():135] Probing git +2024-05-14 16:34:24,963 ERROR HandlerThread:117696 [gitlib.py:root():92] git root error: Cmd('git') failed due to: exit code(128) + cmdline: git rev-parse --show-toplevel + stderr: 'fatal: detected dubious ownership in repository at '/data/cronscript/lm-evaluation-harness' +To add an exception for this directory, call: + + git config --global --add safe.directory /data/cronscript/lm-evaluation-harness' +2024-05-14 16:34:24,963 DEBUG HandlerThread:117696 [system_info.py:_probe_git():143] Probing git done +2024-05-14 16:34:24,963 DEBUG HandlerThread:117696 [system_info.py:probe():198] Probing system done +2024-05-14 16:34:24,963 DEBUG HandlerThread:117696 [system_monitor.py:probe():223] {'os': 'Linux-5.15.0-92-generic-x86_64-with-glibc2.35', 'python': '3.10.12', 'heartbeatAt': '2024-05-14T16:34:24.936081', 'startedAt': '2024-05-14T16:34:24.428958', 'docker': None, 'cuda': None, 'args': ('--model', 'hf', '--model_args', 'pretrained=/data/cronscript/ckpts//hf_ckpt//global_step100', '--tasks', 'indiccopa-hi', '--batch_size', 'auto', '--wandb_args', 'project=bharatgpt,group=trial_expt'), 'state': 'running', 'program': '-m lm_eval.__main__', 'codePathLocal': None, 'git': {'remote': 'https://github.com/EleutherAI/lm-evaluation-harness', 'commit': None}, 'email': None, 'root': '/data/cronscript/lm-evaluation-harness', 'host': 'vizzhy-150-3', 'username': 'root', 'executable': '/usr/bin/python3', 'cpu_count': 76, 'cpu_count_logical': 152, 'cpu_freq': {'current': 3394.7368157894734, 'min': 800.0, 'max': 3400.0}, 'cpu_freq_per_core': [{'current': 3300.0, 'min': 800.0, 'max': 3400.0}, {'current': 3300.0, 'min': 800.0, 'max': 3400.0}, {'current': 3300.002, 'min': 800.0, 'max': 3400.0}, {'current': 3300.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3313.888, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3210.166, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}], 'disk': {'/': {'total': 866.4415092468262, 'used': 863.4841461181641}}, 'memory': {'total': 1007.5000267028809}} +2024-05-14 16:34:24,964 INFO HandlerThread:117696 [system_monitor.py:probe():224] Finished collecting system info +2024-05-14 16:34:24,964 INFO HandlerThread:117696 [system_monitor.py:probe():227] Publishing system info +2024-05-14 16:34:24,965 INFO HandlerThread:117696 [system_monitor.py:probe():229] Finished publishing system info +2024-05-14 16:34:24,969 DEBUG SenderThread:117696 [sender.py:send():378] send: files +2024-05-14 16:34:24,969 INFO SenderThread:117696 [sender.py:_save_file():1389] saving file wandb-metadata.json with policy now +2024-05-14 16:34:25,067 DEBUG HandlerThread:117696 [handler.py:handle_request():158] handle_request: python_packages +2024-05-14 16:34:25,067 DEBUG SenderThread:117696 [sender.py:send_request():405] send_request: python_packages +2024-05-14 16:34:25,068 DEBUG SenderThread:117696 [sender.py:send():378] send: telemetry +2024-05-14 16:34:25,069 DEBUG HandlerThread:117696 [handler.py:handle_request():158] handle_request: stop_status +2024-05-14 16:34:25,069 DEBUG SenderThread:117696 [sender.py:send_request():405] send_request: stop_status +2024-05-14 16:34:25,508 INFO wandb-upload_0:117696 [upload_job.py:push():130] Uploaded file /tmp/tmpkoay5zncwandb/abvzid9c-wandb-metadata.json +2024-05-14 16:34:25,755 INFO Thread-12 :117696 [dir_watcher.py:_on_file_created():271] file/dir created: /data/cronscript/lm-evaluation-harness/wandb/run-20240514_163424-53eroz7t/files/output.log +2024-05-14 16:34:25,755 INFO Thread-12 :117696 [dir_watcher.py:_on_file_created():271] file/dir created: /data/cronscript/lm-evaluation-harness/wandb/run-20240514_163424-53eroz7t/files/requirements.txt +2024-05-14 16:34:25,755 INFO Thread-12 :117696 [dir_watcher.py:_on_file_created():271] file/dir created: /data/cronscript/lm-evaluation-harness/wandb/run-20240514_163424-53eroz7t/files/wandb-metadata.json +2024-05-14 16:34:27,755 INFO Thread-12 :117696 [dir_watcher.py:_on_file_modified():288] file/dir modified: /data/cronscript/lm-evaluation-harness/wandb/run-20240514_163424-53eroz7t/files/output.log +2024-05-14 16:34:30,312 DEBUG HandlerThread:117696 [handler.py:handle_request():158] handle_request: status_report +2024-05-14 16:34:31,758 INFO Thread-12 :117696 [dir_watcher.py:_on_file_modified():288] file/dir modified: /data/cronscript/lm-evaluation-harness/wandb/run-20240514_163424-53eroz7t/files/output.log +2024-05-14 16:34:35,623 DEBUG HandlerThread:117696 [handler.py:handle_request():158] handle_request: status_report +2024-05-14 16:34:39,770 INFO Thread-12 :117696 [dir_watcher.py:_on_file_modified():288] file/dir modified: /data/cronscript/lm-evaluation-harness/wandb/run-20240514_163424-53eroz7t/files/output.log +2024-05-14 16:34:40,068 DEBUG HandlerThread:117696 [handler.py:handle_request():158] handle_request: stop_status +2024-05-14 16:34:40,069 DEBUG SenderThread:117696 [sender.py:send_request():405] send_request: stop_status +2024-05-14 16:34:41,152 DEBUG HandlerThread:117696 [handler.py:handle_request():158] handle_request: status_report +2024-05-14 16:34:43,774 INFO Thread-12 :117696 [dir_watcher.py:_on_file_modified():288] file/dir modified: /data/cronscript/lm-evaluation-harness/wandb/run-20240514_163424-53eroz7t/files/output.log +2024-05-14 16:34:45,775 INFO Thread-12 :117696 [dir_watcher.py:_on_file_modified():288] file/dir modified: /data/cronscript/lm-evaluation-harness/wandb/run-20240514_163424-53eroz7t/files/output.log +2024-05-14 16:34:46,280 DEBUG HandlerThread:117696 [handler.py:handle_request():158] handle_request: status_report +2024-05-14 16:34:47,777 INFO Thread-12 :117696 [dir_watcher.py:_on_file_modified():288] file/dir modified: /data/cronscript/lm-evaluation-harness/wandb/run-20240514_163424-53eroz7t/files/output.log +2024-05-14 16:34:51,281 DEBUG HandlerThread:117696 [handler.py:handle_request():158] handle_request: status_report +2024-05-14 16:34:52,296 DEBUG SenderThread:117696 [sender.py:send():378] send: exit +2024-05-14 16:34:52,296 INFO SenderThread:117696 [sender.py:send_exit():585] handling exit code: 0 +2024-05-14 16:34:52,296 INFO SenderThread:117696 [sender.py:send_exit():587] handling runtime: 27 +2024-05-14 16:34:52,298 INFO SenderThread:117696 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end +2024-05-14 16:34:52,298 INFO SenderThread:117696 [sender.py:send_exit():593] send defer +2024-05-14 16:34:52,298 DEBUG HandlerThread:117696 [handler.py:handle_request():158] handle_request: defer +2024-05-14 16:34:52,298 INFO HandlerThread:117696 [handler.py:handle_request_defer():184] handle defer: 0 +2024-05-14 16:34:52,298 DEBUG SenderThread:117696 [sender.py:send_request():405] send_request: defer +2024-05-14 16:34:52,298 INFO SenderThread:117696 [sender.py:send_request_defer():609] handle sender defer: 0 +2024-05-14 16:34:52,298 INFO SenderThread:117696 [sender.py:transition_state():613] send defer: 1 +2024-05-14 16:34:52,299 DEBUG HandlerThread:117696 [handler.py:handle_request():158] handle_request: defer +2024-05-14 16:34:52,299 INFO HandlerThread:117696 [handler.py:handle_request_defer():184] handle defer: 1 +2024-05-14 16:34:52,299 DEBUG SenderThread:117696 [sender.py:send_request():405] send_request: defer +2024-05-14 16:34:52,299 INFO SenderThread:117696 [sender.py:send_request_defer():609] handle sender defer: 1 +2024-05-14 16:34:52,299 INFO SenderThread:117696 [sender.py:transition_state():613] send defer: 2 +2024-05-14 16:34:52,299 DEBUG HandlerThread:117696 [handler.py:handle_request():158] handle_request: defer +2024-05-14 16:34:52,299 INFO HandlerThread:117696 [handler.py:handle_request_defer():184] handle defer: 2 +2024-05-14 16:34:52,299 INFO HandlerThread:117696 [system_monitor.py:finish():203] Stopping system monitor +2024-05-14 16:34:52,299 DEBUG SystemMonitor:117696 [system_monitor.py:_start():172] Starting system metrics aggregation loop +2024-05-14 16:34:52,300 INFO HandlerThread:117696 [interfaces.py:finish():200] Joined cpu monitor +2024-05-14 16:34:52,300 DEBUG SystemMonitor:117696 [system_monitor.py:_start():179] Finished system metrics aggregation loop +2024-05-14 16:34:52,300 INFO HandlerThread:117696 [interfaces.py:finish():200] Joined disk monitor +2024-05-14 16:34:52,300 DEBUG SystemMonitor:117696 [system_monitor.py:_start():183] Publishing last batch of metrics +2024-05-14 16:34:52,300 INFO HandlerThread:117696 [interfaces.py:finish():200] Joined memory monitor +2024-05-14 16:34:52,303 INFO HandlerThread:117696 [interfaces.py:finish():200] Joined network monitor +2024-05-14 16:34:52,303 DEBUG SenderThread:117696 [sender.py:send_request():405] send_request: defer +2024-05-14 16:34:52,303 INFO SenderThread:117696 [sender.py:send_request_defer():609] handle sender defer: 2 +2024-05-14 16:34:52,303 INFO SenderThread:117696 [sender.py:transition_state():613] send defer: 3 +2024-05-14 16:34:52,303 DEBUG HandlerThread:117696 [handler.py:handle_request():158] handle_request: defer +2024-05-14 16:34:52,303 INFO HandlerThread:117696 [handler.py:handle_request_defer():184] handle defer: 3 +2024-05-14 16:34:52,304 DEBUG SenderThread:117696 [sender.py:send():378] send: stats +2024-05-14 16:34:52,304 DEBUG SenderThread:117696 [sender.py:send_request():405] send_request: defer +2024-05-14 16:34:52,304 INFO SenderThread:117696 [sender.py:send_request_defer():609] handle sender defer: 3 +2024-05-14 16:34:52,304 INFO SenderThread:117696 [sender.py:transition_state():613] send defer: 4 +2024-05-14 16:34:52,304 DEBUG HandlerThread:117696 [handler.py:handle_request():158] handle_request: defer +2024-05-14 16:34:52,304 INFO HandlerThread:117696 [handler.py:handle_request_defer():184] handle defer: 4 +2024-05-14 16:34:52,305 DEBUG SenderThread:117696 [sender.py:send_request():405] send_request: defer +2024-05-14 16:34:52,305 INFO SenderThread:117696 [sender.py:send_request_defer():609] handle sender defer: 4 +2024-05-14 16:34:52,305 INFO SenderThread:117696 [sender.py:transition_state():613] send defer: 5 +2024-05-14 16:34:52,305 DEBUG HandlerThread:117696 [handler.py:handle_request():158] handle_request: defer +2024-05-14 16:34:52,305 INFO HandlerThread:117696 [handler.py:handle_request_defer():184] handle defer: 5 +2024-05-14 16:34:52,305 DEBUG SenderThread:117696 [sender.py:send():378] send: summary +2024-05-14 16:34:52,306 INFO SenderThread:117696 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end +2024-05-14 16:34:52,306 DEBUG SenderThread:117696 [sender.py:send_request():405] send_request: defer +2024-05-14 16:34:52,306 INFO SenderThread:117696 [sender.py:send_request_defer():609] handle sender defer: 5 +2024-05-14 16:34:52,306 INFO SenderThread:117696 [sender.py:transition_state():613] send defer: 6 +2024-05-14 16:34:52,306 DEBUG HandlerThread:117696 [handler.py:handle_request():158] handle_request: defer +2024-05-14 16:34:52,306 INFO HandlerThread:117696 [handler.py:handle_request_defer():184] handle defer: 6 +2024-05-14 16:34:52,306 DEBUG SenderThread:117696 [sender.py:send_request():405] send_request: defer +2024-05-14 16:34:52,306 INFO SenderThread:117696 [sender.py:send_request_defer():609] handle sender defer: 6 +2024-05-14 16:34:52,309 DEBUG HandlerThread:117696 [handler.py:handle_request():158] handle_request: status_report +2024-05-14 16:34:52,393 INFO SenderThread:117696 [sender.py:transition_state():613] send defer: 7 +2024-05-14 16:34:52,393 DEBUG HandlerThread:117696 [handler.py:handle_request():158] handle_request: defer +2024-05-14 16:34:52,393 INFO HandlerThread:117696 [handler.py:handle_request_defer():184] handle defer: 7 +2024-05-14 16:34:52,393 DEBUG SenderThread:117696 [sender.py:send_request():405] send_request: defer +2024-05-14 16:34:52,394 INFO SenderThread:117696 [sender.py:send_request_defer():609] handle sender defer: 7 +2024-05-14 16:34:52,780 INFO Thread-12 :117696 [dir_watcher.py:_on_file_modified():288] file/dir modified: /data/cronscript/lm-evaluation-harness/wandb/run-20240514_163424-53eroz7t/files/config.yaml +2024-05-14 16:34:52,780 INFO Thread-12 :117696 [dir_watcher.py:_on_file_created():271] file/dir created: /data/cronscript/lm-evaluation-harness/wandb/run-20240514_163424-53eroz7t/files/wandb-summary.json +2024-05-14 16:34:53,296 DEBUG HandlerThread:117696 [handler.py:handle_request():158] handle_request: poll_exit +2024-05-14 16:34:55,235 INFO SenderThread:117696 [sender.py:transition_state():613] send defer: 8 +2024-05-14 16:34:55,235 DEBUG SenderThread:117696 [sender.py:send_request():405] send_request: poll_exit +2024-05-14 16:34:55,235 DEBUG HandlerThread:117696 [handler.py:handle_request():158] handle_request: defer +2024-05-14 16:34:55,236 INFO HandlerThread:117696 [handler.py:handle_request_defer():184] handle defer: 8 +2024-05-14 16:34:55,236 DEBUG SenderThread:117696 [sender.py:send_request():405] send_request: defer +2024-05-14 16:34:55,236 INFO SenderThread:117696 [sender.py:send_request_defer():609] handle sender defer: 8 +2024-05-14 16:34:55,236 INFO SenderThread:117696 [job_builder.py:build():432] Attempting to build job artifact +2024-05-14 16:34:55,236 INFO SenderThread:117696 [job_builder.py:_get_source_type():576] no source found +2024-05-14 16:34:55,236 INFO SenderThread:117696 [sender.py:transition_state():613] send defer: 9 +2024-05-14 16:34:55,236 DEBUG HandlerThread:117696 [handler.py:handle_request():158] handle_request: defer +2024-05-14 16:34:55,237 INFO HandlerThread:117696 [handler.py:handle_request_defer():184] handle defer: 9 +2024-05-14 16:34:55,237 DEBUG SenderThread:117696 [sender.py:send_request():405] send_request: defer +2024-05-14 16:34:55,237 INFO SenderThread:117696 [sender.py:send_request_defer():609] handle sender defer: 9 +2024-05-14 16:34:55,237 INFO SenderThread:117696 [dir_watcher.py:finish():358] shutting down directory watcher +2024-05-14 16:34:55,296 DEBUG HandlerThread:117696 [handler.py:handle_request():158] handle_request: poll_exit +2024-05-14 16:34:55,781 INFO SenderThread:117696 [dir_watcher.py:_on_file_modified():288] file/dir modified: /data/cronscript/lm-evaluation-harness/wandb/run-20240514_163424-53eroz7t/files/output.log +2024-05-14 16:34:55,782 INFO SenderThread:117696 [dir_watcher.py:finish():388] scan: /data/cronscript/lm-evaluation-harness/wandb/run-20240514_163424-53eroz7t/files +2024-05-14 16:34:55,782 INFO SenderThread:117696 [dir_watcher.py:finish():402] scan save: /data/cronscript/lm-evaluation-harness/wandb/run-20240514_163424-53eroz7t/files/config.yaml config.yaml +2024-05-14 16:34:55,782 INFO SenderThread:117696 [dir_watcher.py:finish():402] scan save: /data/cronscript/lm-evaluation-harness/wandb/run-20240514_163424-53eroz7t/files/wandb-summary.json wandb-summary.json +2024-05-14 16:34:55,782 INFO SenderThread:117696 [dir_watcher.py:finish():402] scan save: /data/cronscript/lm-evaluation-harness/wandb/run-20240514_163424-53eroz7t/files/output.log output.log +2024-05-14 16:34:55,782 INFO SenderThread:117696 [dir_watcher.py:finish():402] scan save: /data/cronscript/lm-evaluation-harness/wandb/run-20240514_163424-53eroz7t/files/requirements.txt requirements.txt +2024-05-14 16:34:55,783 INFO SenderThread:117696 [dir_watcher.py:finish():402] scan save: /data/cronscript/lm-evaluation-harness/wandb/run-20240514_163424-53eroz7t/files/wandb-metadata.json wandb-metadata.json +2024-05-14 16:34:55,783 INFO SenderThread:117696 [sender.py:transition_state():613] send defer: 10 +2024-05-14 16:34:55,783 DEBUG SenderThread:117696 [sender.py:send_request():405] send_request: poll_exit +2024-05-14 16:34:55,783 DEBUG HandlerThread:117696 [handler.py:handle_request():158] handle_request: defer +2024-05-14 16:34:55,786 INFO HandlerThread:117696 [handler.py:handle_request_defer():184] handle defer: 10 +2024-05-14 16:34:55,787 DEBUG SenderThread:117696 [sender.py:send_request():405] send_request: defer +2024-05-14 16:34:55,787 INFO SenderThread:117696 [sender.py:send_request_defer():609] handle sender defer: 10 +2024-05-14 16:34:55,787 INFO SenderThread:117696 [file_pusher.py:finish():169] shutting down file pusher +2024-05-14 16:34:56,034 INFO wandb-upload_1:117696 [upload_job.py:push():130] Uploaded file /data/cronscript/lm-evaluation-harness/wandb/run-20240514_163424-53eroz7t/files/wandb-summary.json +2024-05-14 16:34:56,191 INFO wandb-upload_0:117696 [upload_job.py:push():130] Uploaded file /data/cronscript/lm-evaluation-harness/wandb/run-20240514_163424-53eroz7t/files/config.yaml +2024-05-14 16:34:56,281 INFO wandb-upload_3:117696 [upload_job.py:push():130] Uploaded file /data/cronscript/lm-evaluation-harness/wandb/run-20240514_163424-53eroz7t/files/requirements.txt +2024-05-14 16:34:56,286 INFO wandb-upload_2:117696 [upload_job.py:push():130] Uploaded file /data/cronscript/lm-evaluation-harness/wandb/run-20240514_163424-53eroz7t/files/output.log +2024-05-14 16:34:56,297 DEBUG HandlerThread:117696 [handler.py:handle_request():158] handle_request: poll_exit +2024-05-14 16:34:56,297 DEBUG SenderThread:117696 [sender.py:send_request():405] send_request: poll_exit +2024-05-14 16:34:56,486 INFO Thread-11 (_thread_body):117696 [sender.py:transition_state():613] send defer: 11 +2024-05-14 16:34:56,487 DEBUG HandlerThread:117696 [handler.py:handle_request():158] handle_request: defer +2024-05-14 16:34:56,487 INFO HandlerThread:117696 [handler.py:handle_request_defer():184] handle defer: 11 +2024-05-14 16:34:56,487 DEBUG SenderThread:117696 [sender.py:send_request():405] send_request: defer +2024-05-14 16:34:56,487 INFO SenderThread:117696 [sender.py:send_request_defer():609] handle sender defer: 11 +2024-05-14 16:34:56,487 INFO SenderThread:117696 [file_pusher.py:join():175] waiting for file pusher +2024-05-14 16:34:56,487 INFO SenderThread:117696 [sender.py:transition_state():613] send defer: 12 +2024-05-14 16:34:56,488 DEBUG HandlerThread:117696 [handler.py:handle_request():158] handle_request: defer +2024-05-14 16:34:56,488 INFO HandlerThread:117696 [handler.py:handle_request_defer():184] handle defer: 12 +2024-05-14 16:34:56,488 DEBUG SenderThread:117696 [sender.py:send_request():405] send_request: defer +2024-05-14 16:34:56,488 INFO SenderThread:117696 [sender.py:send_request_defer():609] handle sender defer: 12 +2024-05-14 16:34:56,488 INFO SenderThread:117696 [file_stream.py:finish():601] file stream finish called +2024-05-14 16:34:56,549 INFO SenderThread:117696 [file_stream.py:finish():605] file stream finish is done +2024-05-14 16:34:56,549 INFO SenderThread:117696 [sender.py:transition_state():613] send defer: 13 +2024-05-14 16:34:56,549 DEBUG HandlerThread:117696 [handler.py:handle_request():158] handle_request: defer +2024-05-14 16:34:56,549 INFO HandlerThread:117696 [handler.py:handle_request_defer():184] handle defer: 13 +2024-05-14 16:34:56,549 DEBUG SenderThread:117696 [sender.py:send_request():405] send_request: defer +2024-05-14 16:34:56,549 INFO SenderThread:117696 [sender.py:send_request_defer():609] handle sender defer: 13 +2024-05-14 16:34:56,549 INFO SenderThread:117696 [sender.py:transition_state():613] send defer: 14 +2024-05-14 16:34:56,549 DEBUG HandlerThread:117696 [handler.py:handle_request():158] handle_request: defer +2024-05-14 16:34:56,550 INFO HandlerThread:117696 [handler.py:handle_request_defer():184] handle defer: 14 +2024-05-14 16:34:56,550 DEBUG SenderThread:117696 [sender.py:send():378] send: final +2024-05-14 16:34:56,550 DEBUG SenderThread:117696 [sender.py:send():378] send: footer +2024-05-14 16:34:56,550 DEBUG SenderThread:117696 [sender.py:send_request():405] send_request: defer +2024-05-14 16:34:56,550 INFO SenderThread:117696 [sender.py:send_request_defer():609] handle sender defer: 14 +2024-05-14 16:34:56,550 DEBUG HandlerThread:117696 [handler.py:handle_request():158] handle_request: poll_exit +2024-05-14 16:34:56,550 DEBUG SenderThread:117696 [sender.py:send_request():405] send_request: poll_exit +2024-05-14 16:34:56,550 DEBUG HandlerThread:117696 [handler.py:handle_request():158] handle_request: poll_exit +2024-05-14 16:34:56,551 DEBUG SenderThread:117696 [sender.py:send_request():405] send_request: poll_exit +2024-05-14 16:34:56,551 DEBUG HandlerThread:117696 [handler.py:handle_request():158] handle_request: server_info +2024-05-14 16:34:56,551 DEBUG SenderThread:117696 [sender.py:send_request():405] send_request: server_info +2024-05-14 16:34:56,552 DEBUG HandlerThread:117696 [handler.py:handle_request():158] handle_request: get_summary +2024-05-14 16:34:56,552 DEBUG HandlerThread:117696 [handler.py:handle_request():158] handle_request: sampled_history +2024-05-14 16:34:56,552 DEBUG HandlerThread:117696 [handler.py:handle_request():158] handle_request: internal_messages +2024-05-14 16:34:56,612 INFO MainThread:117696 [wandb_run.py:_footer_history_summary_info():3994] rendering history +2024-05-14 16:34:56,612 INFO MainThread:117696 [wandb_run.py:_footer_history_summary_info():4026] rendering summary +2024-05-14 16:34:56,612 INFO MainThread:117696 [wandb_run.py:_footer_sync_info():3953] logging synced files +2024-05-14 16:34:56,612 DEBUG HandlerThread:117696 [handler.py:handle_request():158] handle_request: shutdown +2024-05-14 16:34:56,612 INFO HandlerThread:117696 [handler.py:finish():882] shutting down handler +2024-05-14 16:34:57,551 INFO WriterThread:117696 [datastore.py:close():296] close: /data/cronscript/lm-evaluation-harness/wandb/run-20240514_163424-53eroz7t/run-53eroz7t.wandb +2024-05-14 16:34:57,612 INFO SenderThread:117696 [sender.py:finish():1545] shutting down sender +2024-05-14 16:34:57,612 INFO SenderThread:117696 [file_pusher.py:finish():169] shutting down file pusher +2024-05-14 16:34:57,612 INFO SenderThread:117696 [file_pusher.py:join():175] waiting for file pusher diff --git a/lm-evaluation-harness/wandb/run-20240514_163424-53eroz7t/logs/debug.log b/lm-evaluation-harness/wandb/run-20240514_163424-53eroz7t/logs/debug.log new file mode 100644 index 0000000000000000000000000000000000000000..1482d8267f04e8394775458d55f3d59dee214a14 --- /dev/null +++ b/lm-evaluation-harness/wandb/run-20240514_163424-53eroz7t/logs/debug.log @@ -0,0 +1,29 @@ +2024-05-14 16:34:24,436 INFO MainThread:116793 [wandb_setup.py:_flush():76] Current SDK version is 0.17.0 +2024-05-14 16:34:24,436 INFO MainThread:116793 [wandb_setup.py:_flush():76] Configure stats pid to 116793 +2024-05-14 16:34:24,436 INFO MainThread:116793 [wandb_setup.py:_flush():76] Loading settings from /root/.config/wandb/settings +2024-05-14 16:34:24,436 INFO MainThread:116793 [wandb_setup.py:_flush():76] Loading settings from /data/cronscript/lm-evaluation-harness/wandb/settings +2024-05-14 16:34:24,436 INFO MainThread:116793 [wandb_setup.py:_flush():76] Loading settings from environment variables: {} +2024-05-14 16:34:24,436 INFO MainThread:116793 [wandb_setup.py:_flush():76] Applying setup settings: {'_disable_service': False} +2024-05-14 16:34:24,436 WARNING MainThread:116793 [wandb_setup.py:_flush():76] Could not find program at -m lm_eval.__main__ +2024-05-14 16:34:24,436 INFO MainThread:116793 [wandb_setup.py:_flush():76] Inferring run settings from compute environment: {'program_relpath': None, 'program': '-m lm_eval.__main__'} +2024-05-14 16:34:24,436 INFO MainThread:116793 [wandb_setup.py:_flush():76] Applying login settings: {} +2024-05-14 16:34:24,436 INFO MainThread:116793 [wandb_init.py:_log_setup():520] Logging user logs to /data/cronscript/lm-evaluation-harness/wandb/run-20240514_163424-53eroz7t/logs/debug.log +2024-05-14 16:34:24,436 INFO MainThread:116793 [wandb_init.py:_log_setup():521] Logging internal logs to /data/cronscript/lm-evaluation-harness/wandb/run-20240514_163424-53eroz7t/logs/debug-internal.log +2024-05-14 16:34:24,436 INFO MainThread:116793 [wandb_init.py:init():560] calling init triggers +2024-05-14 16:34:24,436 INFO MainThread:116793 [wandb_init.py:init():567] wandb.init called with sweep_config: {} +config: {} +2024-05-14 16:34:24,436 INFO MainThread:116793 [wandb_init.py:init():610] starting backend +2024-05-14 16:34:24,436 INFO MainThread:116793 [wandb_init.py:init():614] setting up manager +2024-05-14 16:34:24,438 INFO MainThread:116793 [backend.py:_multiprocessing_setup():105] multiprocessing start_methods=fork,spawn,forkserver, using: spawn +2024-05-14 16:34:24,439 INFO MainThread:116793 [wandb_init.py:init():622] backend started and connected +2024-05-14 16:34:24,442 INFO MainThread:116793 [wandb_init.py:init():711] updated telemetry +2024-05-14 16:34:24,453 INFO MainThread:116793 [wandb_init.py:init():744] communicating run to backend with 90.0 second timeout +2024-05-14 16:34:24,758 INFO MainThread:116793 [wandb_run.py:_on_init():2396] communicating current version +2024-05-14 16:34:24,843 INFO MainThread:116793 [wandb_run.py:_on_init():2405] got version response +2024-05-14 16:34:24,843 INFO MainThread:116793 [wandb_init.py:init():795] starting run threads in backend +2024-05-14 16:34:25,067 INFO MainThread:116793 [wandb_run.py:_console_start():2374] atexit reg +2024-05-14 16:34:25,067 INFO MainThread:116793 [wandb_run.py:_redirect():2229] redirect: wrap_raw +2024-05-14 16:34:25,068 INFO MainThread:116793 [wandb_run.py:_redirect():2294] Wrapping output streams. +2024-05-14 16:34:25,068 INFO MainThread:116793 [wandb_run.py:_redirect():2319] Redirects installed. +2024-05-14 16:34:25,069 INFO MainThread:116793 [wandb_init.py:init():838] run started, returning control to user process +2024-05-14 16:34:57,613 WARNING MsgRouterThr:116793 [router.py:message_loop():77] message_loop has been closed diff --git a/lm-evaluation-harness/wandb/run-20240514_163424-53eroz7t/run-53eroz7t.wandb b/lm-evaluation-harness/wandb/run-20240514_163424-53eroz7t/run-53eroz7t.wandb new file mode 100644 index 0000000000000000000000000000000000000000..890843cb4b0a13e2f35a26c2610e11103a526d0e Binary files /dev/null and b/lm-evaluation-harness/wandb/run-20240514_163424-53eroz7t/run-53eroz7t.wandb differ diff --git a/lm-evaluation-harness/wandb/run-20240522_184737-460fnitv/files/config.yaml b/lm-evaluation-harness/wandb/run-20240522_184737-460fnitv/files/config.yaml new file mode 100644 index 0000000000000000000000000000000000000000..143ad1480a78437631e227487194d265237c88de --- /dev/null +++ b/lm-evaluation-harness/wandb/run-20240522_184737-460fnitv/files/config.yaml @@ -0,0 +1,43 @@ +wandb_version: 1 + +_wandb: + desc: null + value: + python_version: 3.10.12 + cli_version: 0.17.0 + framework: huggingface + huggingface_version: 4.41.0 + is_jupyter_run: false + is_kaggle_kernel: false + start_time: 1716403657 + t: + 1: + - 1 + - 5 + - 11 + - 49 + - 51 + - 53 + - 55 + - 71 + - 98 + - 100 + 2: + - 1 + - 5 + - 11 + - 49 + - 51 + - 53 + - 55 + - 71 + - 98 + - 100 + 3: + - 23 + 4: 3.10.12 + 5: 0.17.0 + 6: 4.41.0 + 8: + - 5 + 13: linux-x86_64 diff --git a/lm-evaluation-harness/wandb/run-20240522_184737-460fnitv/files/requirements.txt b/lm-evaluation-harness/wandb/run-20240522_184737-460fnitv/files/requirements.txt new file mode 100644 index 0000000000000000000000000000000000000000..a3573b3bbfd5b190fed4ccaed4ac2846002aec22 --- /dev/null +++ b/lm-evaluation-harness/wandb/run-20240522_184737-460fnitv/files/requirements.txt @@ -0,0 +1,155 @@ +DataProperty==1.0.1 +GitPython==3.1.43 +Jinja2==3.1.4 +Markdown==3.6 +MarkupSafe==2.1.5 +Pillow-SIMD==7.0.0.post3 +PyYAML==6.0 +Werkzeug==3.0.3 +absl-py==2.1.0 +accelerate==0.30.1 +aiohttp==3.9.5 +aiosignal==1.3.1 +async-timeout==4.0.3 +attrs==23.2.0 +av==9.2.0 +cachetools==5.3.3 +certifi==2024.2.2 +cffi==1.15.1 +cfgv==3.4.0 +chardet==5.2.0 +charset-normalizer==3.3.2 +click==8.1.7 +cmake==3.29.2 +colorama==0.4.6 +datasets==2.19.1 +deepspeed==0.12.4+hpu.synapse.v1.15.1 +dill==0.3.8 +distlib==0.3.8 +docker-pycreds==0.4.0 +einops==0.8.0 +evaluate==0.4.2 +exceptiongroup==1.2.1 +expecttest==0.2.1 +filelock==3.14.0 +frozenlist==1.4.1 +fsspec==2024.3.1 +gitdb==4.0.11 +google-auth-oauthlib==0.4.6 +google-auth==2.29.0 +grpcio==1.63.0 +habana-media-loader==1.15.1.15 +habana-pyhlml==1.15.1.15 +habana-torch-dataloader==1.15.1.15 +habana-torch-plugin==1.15.1.15 +habana_gpu_migration==1.15.1.15 +habana_quantization_toolkit==1.15.1.15 +hjson==3.1.0 +huggingface-hub==0.23.1 +identify==2.5.36 +idna==3.7 +iniconfig==2.0.0 +joblib==1.4.2 +jsonlines==4.0.0 +lightning-habana==1.4.0 +lightning-utilities==0.11.2 +lightning==2.2.0.post0 +lm_eval==0.4.2 +lm_eval==0.4.2 +lm_eval==0.4.2 +lxml==5.2.2 +mbstrdecoder==1.1.3 +more-itertools==10.2.0 +mpi4py==3.1.4 +mpmath==1.3.0 +multidict==6.0.5 +multiprocess==0.70.16 +networkx==3.3 +ninja==1.11.1.1 +nltk==3.8.1 +nodeenv==1.8.0 +numexpr==2.10.0 +numpy==1.23.5 +oauthlib==3.2.2 +packaging==24.0 +pandas==2.0.1 +pathspec==0.12.1 +pathvalidate==3.2.0 +peft==0.11.1 +perfetto==0.7.0 +pillow==10.3.0 +pip==22.0.2 +pip==23.3.1 +platformdirs==4.2.1 +pluggy==1.5.0 +portalocker==2.8.2 +pre-commit==3.3.3 +pretty-errors==1.2.25 +protobuf==3.20.3 +psutil==5.9.8 +py-cpuinfo==9.0.0 +pyarrow-hotfix==0.6 +pyarrow==16.1.0 +pyasn1==0.6.0 +pyasn1_modules==0.4.0 +pybind11==2.10.4 +pycparser==2.22 +pydantic==1.10.13 +pynvml==8.0.4 +pytablewriter==1.2.0 +pytest==8.2.0 +python-dateutil==2.9.0.post0 +pytorch-lightning==2.2.4 +pytz==2024.1 +regex==2023.5.5 +requests-oauthlib==2.0.0 +requests==2.31.0 +rouge_score==0.1.2 +rsa==4.9 +sacrebleu==2.4.2 +safetensors==0.4.3 +scikit-learn==1.5.0 +scipy==1.13.0 +sentencepiece==0.2.0 +sentry-sdk==2.2.1 +setproctitle==1.3.3 +setuptools==59.6.0 +setuptools==69.5.1 +six==1.16.0 +smmap==5.0.1 +sqlitedict==2.1.0 +symengine==0.11.0 +sympy==1.12 +tabledata==1.3.3 +tabulate==0.9.0 +tcolorpy==0.1.6 +tdqm==0.0.1 +tensorboard-data-server==0.6.1 +tensorboard-plugin-wit==1.8.1 +tensorboard==2.11.2 +threadpoolctl==3.5.0 +tokenizers==0.19.1 +tomli==2.0.1 +torch==2.2.0a0+git8964477 +torch_tb_profiler==0.4.0 +torchaudio==2.2.0+08901ad +torchdata==0.7.1+5e6f7b7 +torchmetrics==1.4.0 +torchtext==0.17.0+400da5c +torchvision==0.17.0+b2383d4 +tqdm-multiprocess==0.0.11 +tqdm==4.66.4 +transformers==4.41.0 +typepy==1.3.2 +typing_extensions==4.11.0 +tzdata==2024.1 +urllib3==1.26.18 +virtualenv==20.26.1 +wandb==0.17.0 +wheel==0.37.1 +wheel==0.43.0 +word2number==1.1 +xxhash==3.4.1 +yamllint==1.35.1 +yarl==1.9.4 +zstandard==0.22.0 \ No newline at end of file diff --git a/lm-evaluation-harness/wandb/run-20240522_184737-460fnitv/files/wandb-summary.json b/lm-evaluation-harness/wandb/run-20240522_184737-460fnitv/files/wandb-summary.json new file mode 100644 index 0000000000000000000000000000000000000000..8bf99d152ad35c3699ec8600ecb8b169d4e35875 --- /dev/null +++ b/lm-evaluation-harness/wandb/run-20240522_184737-460fnitv/files/wandb-summary.json @@ -0,0 +1 @@ +{"_wandb": {"runtime": 11}} \ No newline at end of file diff --git a/lm-evaluation-harness/wandb/run-20240522_184737-460fnitv/run-460fnitv.wandb b/lm-evaluation-harness/wandb/run-20240522_184737-460fnitv/run-460fnitv.wandb new file mode 100644 index 0000000000000000000000000000000000000000..c94aaa4a32212c82ca0d896f44ba92ef8c2ca340 Binary files /dev/null and b/lm-evaluation-harness/wandb/run-20240522_184737-460fnitv/run-460fnitv.wandb differ diff --git a/lm-evaluation-harness/wandb/run-20240523_063944-3nhz8pu0/files/config.yaml b/lm-evaluation-harness/wandb/run-20240523_063944-3nhz8pu0/files/config.yaml new file mode 100644 index 0000000000000000000000000000000000000000..5ac0361ac6c9f3d92bcaa27d0084d558d584b56c --- /dev/null +++ b/lm-evaluation-harness/wandb/run-20240523_063944-3nhz8pu0/files/config.yaml @@ -0,0 +1,43 @@ +wandb_version: 1 + +_wandb: + desc: null + value: + python_version: 3.10.12 + cli_version: 0.17.0 + framework: huggingface + huggingface_version: 4.41.1 + is_jupyter_run: false + is_kaggle_kernel: false + start_time: 1716446384 + t: + 1: + - 1 + - 5 + - 11 + - 49 + - 51 + - 53 + - 55 + - 71 + - 98 + - 100 + 2: + - 1 + - 5 + - 11 + - 49 + - 51 + - 53 + - 55 + - 71 + - 98 + - 100 + 3: + - 23 + 4: 3.10.12 + 5: 0.17.0 + 6: 4.41.1 + 8: + - 5 + 13: linux-x86_64 diff --git a/lm-evaluation-harness/wandb/run-20240523_063944-3nhz8pu0/files/output.log b/lm-evaluation-harness/wandb/run-20240523_063944-3nhz8pu0/files/output.log new file mode 100644 index 0000000000000000000000000000000000000000..34aba1b424dc8a53dbaac3b36ead5dc3f58f2951 --- /dev/null +++ b/lm-evaluation-harness/wandb/run-20240523_063944-3nhz8pu0/files/output.log @@ -0,0 +1,34 @@ + +2024-05-23:06:39:45,601 INFO [__main__.py:251] Verbosity set to INFO +2024-05-23:06:39:54,966 INFO [__main__.py:335] Selected Tasks: ['arc_easy', 'hellaswag', 'mrpc', 'openbookqa', 'sst2', 'winogrande'] +2024-05-23:06:39:54,967 INFO [evaluator.py:131] Setting random seed to 0 | Setting numpy seed to 1234 | Setting torch manual seed to 1234 +2024-05-23:06:39:54,969 INFO [evaluator.py:177] Initializing hf model, with arguments: {'pretrained': '/mnt/weka/peacock/experiments/llama/checkpoint/llamav2-3b//hf_ckpt//global_step100'} +2024-05-23:06:39:57,338 INFO [huggingface.py:164] Using device 'cuda' +Traceback (most recent call last): + File "/usr/lib/python3.10/runpy.py", line 196, in _run_module_as_main + return _run_code(code, main_globals, None, + File "/usr/lib/python3.10/runpy.py", line 86, in _run_code + exec(code, run_globals) + File "/mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/lm_eval/__main__.py", line 417, in + cli_evaluate() + File "/mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/lm_eval/__main__.py", line 341, in cli_evaluate + results = evaluator.simple_evaluate( + File "/mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/lm_eval/utils.py", line 288, in _wrapper + return fn(*args, **kwargs) + File "/mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/lm_eval/evaluator.py", line 180, in simple_evaluate + lm = lm_eval.api.registry.get_model(model).create_from_arg_string( + File "/mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/lm_eval/api/model.py", line 134, in create_from_arg_string + return cls(**args, **args2) + File "/mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/lm_eval/models/huggingface.py", line 190, in __init__ + self._get_config( + File "/mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/lm_eval/models/huggingface.py", line 471, in _get_config + self._config = transformers.AutoConfig.from_pretrained( + File "/usr/local/lib/python3.10/dist-packages/transformers/models/auto/configuration_auto.py", line 934, in from_pretrained + config_dict, unused_kwargs = PretrainedConfig.get_config_dict(pretrained_model_name_or_path, **kwargs) + File "/usr/local/lib/python3.10/dist-packages/transformers/configuration_utils.py", line 632, in get_config_dict + config_dict, kwargs = cls._get_config_dict(pretrained_model_name_or_path, **kwargs) + File "/usr/local/lib/python3.10/dist-packages/transformers/configuration_utils.py", line 689, in _get_config_dict + resolved_config_file = cached_file( + File "/usr/local/lib/python3.10/dist-packages/transformers/utils/hub.py", line 370, in cached_file + raise EnvironmentError( +OSError: /mnt/weka/peacock/experiments/llama/checkpoint/llamav2-3b//hf_ckpt//global_step100 does not appear to have a file named config.json. Checkout 'https://huggingface.co//mnt/weka/peacock/experiments/llama/checkpoint/llamav2-3b//hf_ckpt//global_step100/tree/main' for available files. \ No newline at end of file diff --git a/lm-evaluation-harness/wandb/run-20240523_063944-3nhz8pu0/files/requirements.txt b/lm-evaluation-harness/wandb/run-20240523_063944-3nhz8pu0/files/requirements.txt new file mode 100644 index 0000000000000000000000000000000000000000..094efea889e07cd1194f661d9bac8791e99088d6 --- /dev/null +++ b/lm-evaluation-harness/wandb/run-20240523_063944-3nhz8pu0/files/requirements.txt @@ -0,0 +1,154 @@ +DataProperty==1.0.1 +GitPython==3.1.43 +Jinja2==3.1.3 +Markdown==3.6 +MarkupSafe==2.1.5 +Pillow-SIMD==7.0.0.post3 +PyYAML==6.0 +Werkzeug==3.0.2 +absl-py==2.1.0 +accelerate==0.30.1 +aiohttp==3.9.4 +aiosignal==1.3.1 +async-timeout==4.0.3 +attrs==23.2.0 +av==9.2.0 +cachetools==5.3.3 +certifi==2024.2.2 +cffi==1.15.1 +cfgv==3.4.0 +chardet==5.2.0 +charset-normalizer==3.3.2 +click==8.1.7 +cmake==3.29.2 +colorama==0.4.6 +datasets==2.19.1 +deepspeed==0.12.4+hpu.synapse.v1.15.1 +dill==0.3.8 +distlib==0.3.8 +docker-pycreds==0.4.0 +einops==0.8.0 +evaluate==0.4.2 +exceptiongroup==1.2.0 +expecttest==0.2.1 +filelock==3.13.4 +frozenlist==1.4.1 +fsspec==2024.3.1 +gitdb==4.0.11 +google-auth-oauthlib==0.4.6 +google-auth==2.29.0 +grpcio==1.62.1 +habana-media-loader==1.15.1.15 +habana-pyhlml==1.15.1.15 +habana-torch-dataloader==1.15.1.15 +habana-torch-plugin==1.15.1.15 +habana_gpu_migration==1.15.1.15 +habana_quantization_toolkit==1.15.1.15 +hjson==3.1.0 +huggingface-hub==0.23.1 +identify==2.5.35 +idna==3.7 +iniconfig==2.0.0 +joblib==1.4.2 +jsonlines==4.0.0 +lightning-habana==1.4.0 +lightning-utilities==0.11.2 +lightning==2.2.0.post0 +lm_eval==0.4.2 +lm_eval==0.4.2 +lm_eval==0.4.2 +lxml==5.2.2 +mbstrdecoder==1.1.3 +more-itertools==10.2.0 +mpi4py==3.1.4 +mpmath==1.3.0 +multidict==6.0.5 +multiprocess==0.70.16 +networkx==3.3 +ninja==1.11.1.1 +nltk==3.8.1 +nodeenv==1.8.0 +numexpr==2.10.0 +numpy==1.23.5 +oauthlib==3.2.2 +packaging==24.0 +pandas==2.0.1 +pathspec==0.12.1 +pathvalidate==3.2.0 +peft==0.11.1 +perfetto==0.7.0 +pillow==10.3.0 +pip==22.0.2 +pip==23.3.1 +platformdirs==4.2.0 +pluggy==1.4.0 +portalocker==2.8.2 +pre-commit==3.3.3 +protobuf==3.20.3 +psutil==5.9.8 +py-cpuinfo==9.0.0 +pyarrow-hotfix==0.6 +pyarrow==16.1.0 +pyasn1==0.6.0 +pyasn1_modules==0.4.0 +pybind11==2.10.4 +pycparser==2.22 +pydantic==1.10.13 +pynvml==8.0.4 +pytablewriter==1.2.0 +pytest==8.1.1 +python-dateutil==2.9.0.post0 +pytorch-lightning==2.2.2 +pytz==2024.1 +regex==2023.5.5 +requests-oauthlib==2.0.0 +requests==2.31.0 +rouge_score==0.1.2 +rsa==4.9 +sacrebleu==2.4.2 +safetensors==0.4.3 +scikit-learn==1.5.0 +scipy==1.13.1 +sentencepiece==0.2.0 +sentry-sdk==2.2.1 +setproctitle==1.3.3 +setuptools==59.6.0 +setuptools==69.5.1 +six==1.16.0 +smmap==5.0.1 +sqlitedict==2.1.0 +symengine==0.11.0 +sympy==1.12 +tabledata==1.3.3 +tabulate==0.9.0 +tcolorpy==0.1.6 +tdqm==0.0.1 +tensorboard-data-server==0.6.1 +tensorboard-plugin-wit==1.8.1 +tensorboard==2.11.2 +threadpoolctl==3.5.0 +tokenizers==0.19.1 +tomli==2.0.1 +torch==2.2.0a0+git8964477 +torch_tb_profiler==0.4.0 +torchaudio==2.2.0+08901ad +torchdata==0.7.1+5e6f7b7 +torchmetrics==1.3.2 +torchtext==0.17.0+400da5c +torchvision==0.17.0+b2383d4 +tqdm-multiprocess==0.0.11 +tqdm==4.66.2 +transformers==4.41.1 +typepy==1.3.2 +typing_extensions==4.11.0 +tzdata==2024.1 +urllib3==1.26.18 +virtualenv==20.25.1 +wandb==0.17.0 +wheel==0.37.1 +wheel==0.43.0 +word2number==1.1 +xxhash==3.4.1 +yamllint==1.35.1 +yarl==1.9.4 +zstandard==0.22.0 \ No newline at end of file diff --git a/lm-evaluation-harness/wandb/run-20240523_063944-3nhz8pu0/files/wandb-metadata.json b/lm-evaluation-harness/wandb/run-20240523_063944-3nhz8pu0/files/wandb-metadata.json new file mode 100644 index 0000000000000000000000000000000000000000..4e079b2f516d0facf75dc691c4ea37ecb7fa6e0a --- /dev/null +++ b/lm-evaluation-harness/wandb/run-20240523_063944-3nhz8pu0/files/wandb-metadata.json @@ -0,0 +1,850 @@ +{ + "os": "Linux-5.15.0-92-generic-x86_64-with-glibc2.35", + "python": "3.10.12", + "heartbeatAt": "2024-05-23T06:39:45.386782", + "startedAt": "2024-05-23T06:39:44.885687", + "docker": null, + "cuda": null, + "args": [ + "--model", + "hf", + "--model_args", + "pretrained=/mnt/weka/peacock/experiments/llama/checkpoint/llamav2-3b//hf_ckpt//global_step100", + "--tasks", + "hellaswag,arc_easy,openbookqa,winogrande,sst2,mrpc", + "--batch_size", + "auto", + "--wandb_args", + "project=bharatgpt,group=trial_expt_2" + ], + "state": "running", + "program": "-m lm_eval.__main__", + "codePathLocal": null, + "git": { + "remote": "https://github.com/EleutherAI/lm-evaluation-harness", + "commit": null + }, + "email": null, + "root": "/mnt/weka/peacock/idc/cronscript/lm-evaluation-harness", + "host": "peacock-evaluation-worker-0", + "username": "root", + "executable": "/usr/bin/python3", + "cpu_count": 80, + "cpu_count_logical": 160, + "cpu_freq": { + "current": 2334.31985, + "min": 800.0, + "max": 3400.0 + }, + "cpu_freq_per_core": [ + { + "current": 3400.002, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + } + ], + "disk": { + "/": { + "total": 877.6341285705566, + "used": 203.8784294128418 + } + }, + "memory": { + "total": 1007.4379196166992 + } +} diff --git a/lm-evaluation-harness/wandb/run-20240523_063944-3nhz8pu0/files/wandb-summary.json b/lm-evaluation-harness/wandb/run-20240523_063944-3nhz8pu0/files/wandb-summary.json new file mode 100644 index 0000000000000000000000000000000000000000..018b4068e81a342faa2a3b0691dee6965106bcc3 --- /dev/null +++ b/lm-evaluation-harness/wandb/run-20240523_063944-3nhz8pu0/files/wandb-summary.json @@ -0,0 +1 @@ +{"_wandb": {"runtime": 12}} \ No newline at end of file diff --git a/lm-evaluation-harness/wandb/run-20240523_063944-3nhz8pu0/logs/debug-internal.log b/lm-evaluation-harness/wandb/run-20240523_063944-3nhz8pu0/logs/debug-internal.log new file mode 100644 index 0000000000000000000000000000000000000000..f0c7718508e245a2aa37abbe24aad814f55c327c --- /dev/null +++ b/lm-evaluation-harness/wandb/run-20240523_063944-3nhz8pu0/logs/debug-internal.log @@ -0,0 +1,185 @@ +2024-05-23 06:39:44,907 INFO StreamThr :825 [internal.py:wandb_internal():85] W&B internal server running at pid: 825, started at: 2024-05-23 06:39:44.905097 +2024-05-23 06:39:44,912 DEBUG HandlerThread:825 [handler.py:handle_request():158] handle_request: status +2024-05-23 06:39:44,912 INFO WriterThread:825 [datastore.py:open_for_write():87] open: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240523_063944-3nhz8pu0/run-3nhz8pu0.wandb +2024-05-23 06:39:44,915 DEBUG SenderThread:825 [sender.py:send():378] send: header +2024-05-23 06:39:44,919 DEBUG SenderThread:825 [sender.py:send():378] send: run +2024-05-23 06:39:45,186 INFO SenderThread:825 [dir_watcher.py:__init__():211] watching files in: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240523_063944-3nhz8pu0/files +2024-05-23 06:39:45,186 INFO SenderThread:825 [sender.py:_start_run_threads():1123] run started: 3nhz8pu0 with start time 1716446384.906358 +2024-05-23 06:39:45,189 DEBUG HandlerThread:825 [handler.py:handle_request():158] handle_request: check_version +2024-05-23 06:39:45,189 DEBUG SenderThread:825 [sender.py:send_request():405] send_request: check_version +2024-05-23 06:39:45,310 DEBUG HandlerThread:825 [handler.py:handle_request():158] handle_request: run_start +2024-05-23 06:39:45,312 DEBUG HandlerThread:825 [system_info.py:__init__():26] System info init +2024-05-23 06:39:45,312 DEBUG HandlerThread:825 [system_info.py:__init__():41] System info init done +2024-05-23 06:39:45,312 INFO HandlerThread:825 [system_monitor.py:start():194] Starting system monitor +2024-05-23 06:39:45,312 INFO SystemMonitor:825 [system_monitor.py:_start():158] Starting system asset monitoring threads +2024-05-23 06:39:45,312 INFO HandlerThread:825 [system_monitor.py:probe():214] Collecting system info +2024-05-23 06:39:45,320 INFO SystemMonitor:825 [interfaces.py:start():188] Started cpu monitoring +2024-05-23 06:39:45,320 INFO SystemMonitor:825 [interfaces.py:start():188] Started disk monitoring +2024-05-23 06:39:45,326 INFO SystemMonitor:825 [interfaces.py:start():188] Started memory monitoring +2024-05-23 06:39:45,326 INFO SystemMonitor:825 [interfaces.py:start():188] Started network monitoring +2024-05-23 06:39:45,386 DEBUG HandlerThread:825 [system_info.py:probe():150] Probing system +2024-05-23 06:39:45,390 DEBUG HandlerThread:825 [system_info.py:_probe_git():135] Probing git +2024-05-23 06:39:45,400 ERROR HandlerThread:825 [gitlib.py:root():92] git root error: Cmd('git') failed due to: exit code(128) + cmdline: git rev-parse --show-toplevel + stderr: 'fatal: detected dubious ownership in repository at '/mnt/weka/peacock/idc/cronscript/lm-evaluation-harness' +To add an exception for this directory, call: + + git config --global --add safe.directory /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness' +2024-05-23 06:39:45,400 DEBUG HandlerThread:825 [system_info.py:_probe_git():143] Probing git done +2024-05-23 06:39:45,400 DEBUG HandlerThread:825 [system_info.py:probe():198] Probing system done +2024-05-23 06:39:45,400 DEBUG HandlerThread:825 [system_monitor.py:probe():223] {'os': 'Linux-5.15.0-92-generic-x86_64-with-glibc2.35', 'python': '3.10.12', 'heartbeatAt': '2024-05-23T06:39:45.386782', 'startedAt': '2024-05-23T06:39:44.885687', 'docker': None, 'cuda': None, 'args': ('--model', 'hf', '--model_args', 'pretrained=/mnt/weka/peacock/experiments/llama/checkpoint/llamav2-3b//hf_ckpt//global_step100', '--tasks', 'hellaswag,arc_easy,openbookqa,winogrande,sst2,mrpc', '--batch_size', 'auto', '--wandb_args', 'project=bharatgpt,group=trial_expt_2'), 'state': 'running', 'program': '-m lm_eval.__main__', 'codePathLocal': None, 'git': {'remote': 'https://github.com/EleutherAI/lm-evaluation-harness', 'commit': None}, 'email': None, 'root': '/mnt/weka/peacock/idc/cronscript/lm-evaluation-harness', 'host': 'peacock-evaluation-worker-0', 'username': 'root', 'executable': '/usr/bin/python3', 'cpu_count': 80, 'cpu_count_logical': 160, 'cpu_freq': {'current': 2334.31985, 'min': 800.0, 'max': 3400.0}, 'cpu_freq_per_core': [{'current': 3400.002, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}], 'disk': {'/': {'total': 877.6341285705566, 'used': 203.8784294128418}}, 'memory': {'total': 1007.4379196166992}} +2024-05-23 06:39:45,400 INFO HandlerThread:825 [system_monitor.py:probe():224] Finished collecting system info +2024-05-23 06:39:45,400 INFO HandlerThread:825 [system_monitor.py:probe():227] Publishing system info +2024-05-23 06:39:45,404 INFO HandlerThread:825 [system_monitor.py:probe():229] Finished publishing system info +2024-05-23 06:39:45,410 DEBUG SenderThread:825 [sender.py:send():378] send: files +2024-05-23 06:39:45,410 INFO SenderThread:825 [sender.py:_save_file():1389] saving file wandb-metadata.json with policy now +2024-05-23 06:39:45,593 DEBUG HandlerThread:825 [handler.py:handle_request():158] handle_request: python_packages +2024-05-23 06:39:45,593 DEBUG SenderThread:825 [sender.py:send_request():405] send_request: python_packages +2024-05-23 06:39:45,594 DEBUG HandlerThread:825 [handler.py:handle_request():158] handle_request: stop_status +2024-05-23 06:39:45,596 DEBUG SenderThread:825 [sender.py:send_request():405] send_request: stop_status +2024-05-23 06:39:45,725 DEBUG SenderThread:825 [sender.py:send():378] send: telemetry +2024-05-23 06:39:46,031 INFO wandb-upload_0:825 [upload_job.py:push():130] Uploaded file /tmp/tmp5cfn3qhpwandb/621an68z-wandb-metadata.json +2024-05-23 06:39:46,189 INFO Thread-12 :825 [dir_watcher.py:_on_file_created():271] file/dir created: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240523_063944-3nhz8pu0/files/requirements.txt +2024-05-23 06:39:46,189 INFO Thread-12 :825 [dir_watcher.py:_on_file_created():271] file/dir created: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240523_063944-3nhz8pu0/files/wandb-metadata.json +2024-05-23 06:39:46,189 INFO Thread-12 :825 [dir_watcher.py:_on_file_created():271] file/dir created: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240523_063944-3nhz8pu0/files/output.log +2024-05-23 06:39:48,188 INFO Thread-12 :825 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240523_063944-3nhz8pu0/files/output.log +2024-05-23 06:39:50,727 DEBUG HandlerThread:825 [handler.py:handle_request():158] handle_request: status_report +2024-05-23 06:39:55,970 DEBUG HandlerThread:825 [handler.py:handle_request():158] handle_request: status_report +2024-05-23 06:39:56,196 INFO Thread-12 :825 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240523_063944-3nhz8pu0/files/output.log +2024-05-23 06:39:57,351 DEBUG SenderThread:825 [sender.py:send():378] send: exit +2024-05-23 06:39:57,351 INFO SenderThread:825 [sender.py:send_exit():585] handling exit code: 1 +2024-05-23 06:39:57,351 INFO SenderThread:825 [sender.py:send_exit():587] handling runtime: 12 +2024-05-23 06:39:57,352 INFO SenderThread:825 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end +2024-05-23 06:39:57,352 INFO SenderThread:825 [sender.py:send_exit():593] send defer +2024-05-23 06:39:57,355 DEBUG HandlerThread:825 [handler.py:handle_request():158] handle_request: defer +2024-05-23 06:39:57,355 INFO HandlerThread:825 [handler.py:handle_request_defer():184] handle defer: 0 +2024-05-23 06:39:57,355 DEBUG SenderThread:825 [sender.py:send_request():405] send_request: defer +2024-05-23 06:39:57,355 INFO SenderThread:825 [sender.py:send_request_defer():609] handle sender defer: 0 +2024-05-23 06:39:57,355 INFO SenderThread:825 [sender.py:transition_state():613] send defer: 1 +2024-05-23 06:39:57,355 DEBUG HandlerThread:825 [handler.py:handle_request():158] handle_request: defer +2024-05-23 06:39:57,355 INFO HandlerThread:825 [handler.py:handle_request_defer():184] handle defer: 1 +2024-05-23 06:39:57,355 DEBUG SenderThread:825 [sender.py:send_request():405] send_request: defer +2024-05-23 06:39:57,355 INFO SenderThread:825 [sender.py:send_request_defer():609] handle sender defer: 1 +2024-05-23 06:39:57,355 INFO SenderThread:825 [sender.py:transition_state():613] send defer: 2 +2024-05-23 06:39:57,355 DEBUG HandlerThread:825 [handler.py:handle_request():158] handle_request: defer +2024-05-23 06:39:57,355 INFO HandlerThread:825 [handler.py:handle_request_defer():184] handle defer: 2 +2024-05-23 06:39:57,356 INFO HandlerThread:825 [system_monitor.py:finish():203] Stopping system monitor +2024-05-23 06:39:57,356 DEBUG SystemMonitor:825 [system_monitor.py:_start():172] Starting system metrics aggregation loop +2024-05-23 06:39:57,356 DEBUG SystemMonitor:825 [system_monitor.py:_start():179] Finished system metrics aggregation loop +2024-05-23 06:39:57,356 DEBUG SystemMonitor:825 [system_monitor.py:_start():183] Publishing last batch of metrics +2024-05-23 06:39:57,357 INFO HandlerThread:825 [interfaces.py:finish():200] Joined cpu monitor +2024-05-23 06:39:57,357 INFO HandlerThread:825 [interfaces.py:finish():200] Joined disk monitor +2024-05-23 06:39:57,357 INFO HandlerThread:825 [interfaces.py:finish():200] Joined memory monitor +2024-05-23 06:39:57,357 INFO HandlerThread:825 [interfaces.py:finish():200] Joined network monitor +2024-05-23 06:39:57,357 DEBUG SenderThread:825 [sender.py:send_request():405] send_request: defer +2024-05-23 06:39:57,357 INFO SenderThread:825 [sender.py:send_request_defer():609] handle sender defer: 2 +2024-05-23 06:39:57,357 INFO SenderThread:825 [sender.py:transition_state():613] send defer: 3 +2024-05-23 06:39:57,357 DEBUG SenderThread:825 [sender.py:send():378] send: stats +2024-05-23 06:39:57,358 DEBUG HandlerThread:825 [handler.py:handle_request():158] handle_request: defer +2024-05-23 06:39:57,358 INFO HandlerThread:825 [handler.py:handle_request_defer():184] handle defer: 3 +2024-05-23 06:39:57,358 DEBUG SenderThread:825 [sender.py:send_request():405] send_request: defer +2024-05-23 06:39:57,358 INFO SenderThread:825 [sender.py:send_request_defer():609] handle sender defer: 3 +2024-05-23 06:39:57,358 INFO SenderThread:825 [sender.py:transition_state():613] send defer: 4 +2024-05-23 06:39:57,358 DEBUG HandlerThread:825 [handler.py:handle_request():158] handle_request: defer +2024-05-23 06:39:57,358 INFO HandlerThread:825 [handler.py:handle_request_defer():184] handle defer: 4 +2024-05-23 06:39:57,358 DEBUG SenderThread:825 [sender.py:send_request():405] send_request: defer +2024-05-23 06:39:57,358 INFO SenderThread:825 [sender.py:send_request_defer():609] handle sender defer: 4 +2024-05-23 06:39:57,358 INFO SenderThread:825 [sender.py:transition_state():613] send defer: 5 +2024-05-23 06:39:57,358 DEBUG HandlerThread:825 [handler.py:handle_request():158] handle_request: defer +2024-05-23 06:39:57,358 INFO HandlerThread:825 [handler.py:handle_request_defer():184] handle defer: 5 +2024-05-23 06:39:57,358 DEBUG SenderThread:825 [sender.py:send():378] send: summary +2024-05-23 06:39:57,359 INFO SenderThread:825 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end +2024-05-23 06:39:57,359 DEBUG SenderThread:825 [sender.py:send_request():405] send_request: defer +2024-05-23 06:39:57,359 INFO SenderThread:825 [sender.py:send_request_defer():609] handle sender defer: 5 +2024-05-23 06:39:57,359 INFO SenderThread:825 [sender.py:transition_state():613] send defer: 6 +2024-05-23 06:39:57,359 DEBUG HandlerThread:825 [handler.py:handle_request():158] handle_request: defer +2024-05-23 06:39:57,360 INFO HandlerThread:825 [handler.py:handle_request_defer():184] handle defer: 6 +2024-05-23 06:39:57,360 DEBUG SenderThread:825 [sender.py:send_request():405] send_request: defer +2024-05-23 06:39:57,360 INFO SenderThread:825 [sender.py:send_request_defer():609] handle sender defer: 6 +2024-05-23 06:39:57,364 DEBUG HandlerThread:825 [handler.py:handle_request():158] handle_request: status_report +2024-05-23 06:39:57,446 INFO SenderThread:825 [sender.py:transition_state():613] send defer: 7 +2024-05-23 06:39:57,446 DEBUG HandlerThread:825 [handler.py:handle_request():158] handle_request: defer +2024-05-23 06:39:57,446 INFO HandlerThread:825 [handler.py:handle_request_defer():184] handle defer: 7 +2024-05-23 06:39:57,446 DEBUG SenderThread:825 [sender.py:send_request():405] send_request: defer +2024-05-23 06:39:57,446 INFO SenderThread:825 [sender.py:send_request_defer():609] handle sender defer: 7 +2024-05-23 06:39:58,199 INFO Thread-12 :825 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240523_063944-3nhz8pu0/files/config.yaml +2024-05-23 06:39:58,199 INFO Thread-12 :825 [dir_watcher.py:_on_file_created():271] file/dir created: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240523_063944-3nhz8pu0/files/wandb-summary.json +2024-05-23 06:39:58,350 DEBUG HandlerThread:825 [handler.py:handle_request():158] handle_request: poll_exit +2024-05-23 06:39:59,750 INFO SenderThread:825 [sender.py:transition_state():613] send defer: 8 +2024-05-23 06:39:59,751 DEBUG SenderThread:825 [sender.py:send_request():405] send_request: poll_exit +2024-05-23 06:39:59,751 DEBUG HandlerThread:825 [handler.py:handle_request():158] handle_request: defer +2024-05-23 06:39:59,751 INFO HandlerThread:825 [handler.py:handle_request_defer():184] handle defer: 8 +2024-05-23 06:39:59,751 DEBUG SenderThread:825 [sender.py:send_request():405] send_request: defer +2024-05-23 06:39:59,751 INFO SenderThread:825 [sender.py:send_request_defer():609] handle sender defer: 8 +2024-05-23 06:39:59,751 INFO SenderThread:825 [job_builder.py:build():432] Attempting to build job artifact +2024-05-23 06:39:59,752 INFO SenderThread:825 [job_builder.py:_get_source_type():576] no source found +2024-05-23 06:39:59,752 INFO SenderThread:825 [sender.py:transition_state():613] send defer: 9 +2024-05-23 06:39:59,752 DEBUG HandlerThread:825 [handler.py:handle_request():158] handle_request: defer +2024-05-23 06:39:59,752 INFO HandlerThread:825 [handler.py:handle_request_defer():184] handle defer: 9 +2024-05-23 06:39:59,752 DEBUG SenderThread:825 [sender.py:send_request():405] send_request: defer +2024-05-23 06:39:59,752 INFO SenderThread:825 [sender.py:send_request_defer():609] handle sender defer: 9 +2024-05-23 06:39:59,752 INFO SenderThread:825 [dir_watcher.py:finish():358] shutting down directory watcher +2024-05-23 06:40:00,200 INFO SenderThread:825 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240523_063944-3nhz8pu0/files/output.log +2024-05-23 06:40:00,201 INFO SenderThread:825 [dir_watcher.py:finish():388] scan: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240523_063944-3nhz8pu0/files +2024-05-23 06:40:00,201 INFO SenderThread:825 [dir_watcher.py:finish():402] scan save: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240523_063944-3nhz8pu0/files/config.yaml config.yaml +2024-05-23 06:40:00,201 INFO SenderThread:825 [dir_watcher.py:finish():402] scan save: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240523_063944-3nhz8pu0/files/wandb-metadata.json wandb-metadata.json +2024-05-23 06:40:00,204 INFO SenderThread:825 [dir_watcher.py:finish():402] scan save: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240523_063944-3nhz8pu0/files/output.log output.log +2024-05-23 06:40:00,204 INFO SenderThread:825 [dir_watcher.py:finish():402] scan save: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240523_063944-3nhz8pu0/files/requirements.txt requirements.txt +2024-05-23 06:40:00,204 INFO SenderThread:825 [dir_watcher.py:finish():402] scan save: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240523_063944-3nhz8pu0/files/wandb-summary.json wandb-summary.json +2024-05-23 06:40:00,204 INFO SenderThread:825 [sender.py:transition_state():613] send defer: 10 +2024-05-23 06:40:00,204 DEBUG HandlerThread:825 [handler.py:handle_request():158] handle_request: defer +2024-05-23 06:40:00,204 INFO HandlerThread:825 [handler.py:handle_request_defer():184] handle defer: 10 +2024-05-23 06:40:00,204 DEBUG SenderThread:825 [sender.py:send_request():405] send_request: defer +2024-05-23 06:40:00,204 INFO SenderThread:825 [sender.py:send_request_defer():609] handle sender defer: 10 +2024-05-23 06:40:00,204 INFO SenderThread:825 [file_pusher.py:finish():169] shutting down file pusher +2024-05-23 06:40:00,351 DEBUG HandlerThread:825 [handler.py:handle_request():158] handle_request: poll_exit +2024-05-23 06:40:00,351 DEBUG SenderThread:825 [sender.py:send_request():405] send_request: poll_exit +2024-05-23 06:40:00,450 INFO wandb-upload_0:825 [upload_job.py:push():130] Uploaded file /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240523_063944-3nhz8pu0/files/config.yaml +2024-05-23 06:40:00,785 INFO wandb-upload_1:825 [upload_job.py:push():130] Uploaded file /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240523_063944-3nhz8pu0/files/output.log +2024-05-23 06:40:00,842 INFO wandb-upload_2:825 [upload_job.py:push():130] Uploaded file /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240523_063944-3nhz8pu0/files/requirements.txt +2024-05-23 06:40:00,857 INFO wandb-upload_3:825 [upload_job.py:push():130] Uploaded file /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240523_063944-3nhz8pu0/files/wandb-summary.json +2024-05-23 06:40:01,057 INFO Thread-11 (_thread_body):825 [sender.py:transition_state():613] send defer: 11 +2024-05-23 06:40:01,057 DEBUG HandlerThread:825 [handler.py:handle_request():158] handle_request: defer +2024-05-23 06:40:01,057 INFO HandlerThread:825 [handler.py:handle_request_defer():184] handle defer: 11 +2024-05-23 06:40:01,057 DEBUG SenderThread:825 [sender.py:send_request():405] send_request: defer +2024-05-23 06:40:01,057 INFO SenderThread:825 [sender.py:send_request_defer():609] handle sender defer: 11 +2024-05-23 06:40:01,057 INFO SenderThread:825 [file_pusher.py:join():175] waiting for file pusher +2024-05-23 06:40:01,058 INFO SenderThread:825 [sender.py:transition_state():613] send defer: 12 +2024-05-23 06:40:01,058 DEBUG HandlerThread:825 [handler.py:handle_request():158] handle_request: defer +2024-05-23 06:40:01,058 INFO HandlerThread:825 [handler.py:handle_request_defer():184] handle defer: 12 +2024-05-23 06:40:01,058 DEBUG SenderThread:825 [sender.py:send_request():405] send_request: defer +2024-05-23 06:40:01,058 INFO SenderThread:825 [sender.py:send_request_defer():609] handle sender defer: 12 +2024-05-23 06:40:01,058 INFO SenderThread:825 [file_stream.py:finish():601] file stream finish called +2024-05-23 06:40:01,352 DEBUG HandlerThread:825 [handler.py:handle_request():158] handle_request: poll_exit +2024-05-23 06:40:01,909 INFO SenderThread:825 [file_stream.py:finish():605] file stream finish is done +2024-05-23 06:40:01,909 INFO SenderThread:825 [sender.py:transition_state():613] send defer: 13 +2024-05-23 06:40:01,910 DEBUG SenderThread:825 [sender.py:send_request():405] send_request: poll_exit +2024-05-23 06:40:01,910 DEBUG HandlerThread:825 [handler.py:handle_request():158] handle_request: defer +2024-05-23 06:40:01,910 INFO HandlerThread:825 [handler.py:handle_request_defer():184] handle defer: 13 +2024-05-23 06:40:01,910 DEBUG SenderThread:825 [sender.py:send_request():405] send_request: defer +2024-05-23 06:40:01,910 INFO SenderThread:825 [sender.py:send_request_defer():609] handle sender defer: 13 +2024-05-23 06:40:01,910 INFO SenderThread:825 [sender.py:transition_state():613] send defer: 14 +2024-05-23 06:40:01,910 DEBUG HandlerThread:825 [handler.py:handle_request():158] handle_request: defer +2024-05-23 06:40:01,910 INFO HandlerThread:825 [handler.py:handle_request_defer():184] handle defer: 14 +2024-05-23 06:40:01,910 DEBUG SenderThread:825 [sender.py:send():378] send: final +2024-05-23 06:40:01,910 DEBUG SenderThread:825 [sender.py:send():378] send: footer +2024-05-23 06:40:01,911 DEBUG SenderThread:825 [sender.py:send_request():405] send_request: defer +2024-05-23 06:40:01,911 INFO SenderThread:825 [sender.py:send_request_defer():609] handle sender defer: 14 +2024-05-23 06:40:01,911 DEBUG HandlerThread:825 [handler.py:handle_request():158] handle_request: poll_exit +2024-05-23 06:40:01,911 DEBUG SenderThread:825 [sender.py:send_request():405] send_request: poll_exit +2024-05-23 06:40:01,912 DEBUG HandlerThread:825 [handler.py:handle_request():158] handle_request: poll_exit +2024-05-23 06:40:01,912 DEBUG HandlerThread:825 [handler.py:handle_request():158] handle_request: server_info +2024-05-23 06:40:01,912 DEBUG HandlerThread:825 [handler.py:handle_request():158] handle_request: get_summary +2024-05-23 06:40:01,912 DEBUG HandlerThread:825 [handler.py:handle_request():158] handle_request: sampled_history +2024-05-23 06:40:01,912 DEBUG HandlerThread:825 [handler.py:handle_request():158] handle_request: internal_messages +2024-05-23 06:40:01,912 DEBUG SenderThread:825 [sender.py:send_request():405] send_request: poll_exit +2024-05-23 06:40:01,912 DEBUG SenderThread:825 [sender.py:send_request():405] send_request: server_info +2024-05-23 06:40:01,965 INFO MainThread:825 [wandb_run.py:_footer_history_summary_info():3994] rendering history +2024-05-23 06:40:01,965 INFO MainThread:825 [wandb_run.py:_footer_history_summary_info():4026] rendering summary +2024-05-23 06:40:01,966 INFO MainThread:825 [wandb_run.py:_footer_sync_info():3953] logging synced files +2024-05-23 06:40:01,966 DEBUG HandlerThread:825 [handler.py:handle_request():158] handle_request: shutdown +2024-05-23 06:40:01,966 INFO HandlerThread:825 [handler.py:finish():882] shutting down handler +2024-05-23 06:40:02,912 INFO WriterThread:825 [datastore.py:close():296] close: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240523_063944-3nhz8pu0/run-3nhz8pu0.wandb +2024-05-23 06:40:02,965 INFO SenderThread:825 [sender.py:finish():1545] shutting down sender +2024-05-23 06:40:02,965 INFO SenderThread:825 [file_pusher.py:finish():169] shutting down file pusher +2024-05-23 06:40:02,965 INFO SenderThread:825 [file_pusher.py:join():175] waiting for file pusher diff --git a/lm-evaluation-harness/wandb/run-20240523_063944-3nhz8pu0/logs/debug.log b/lm-evaluation-harness/wandb/run-20240523_063944-3nhz8pu0/logs/debug.log new file mode 100644 index 0000000000000000000000000000000000000000..362731eb9c9e13eda64a2d965afa2fd3f919a747 --- /dev/null +++ b/lm-evaluation-harness/wandb/run-20240523_063944-3nhz8pu0/logs/debug.log @@ -0,0 +1,29 @@ +2024-05-23 06:39:44,900 INFO MainThread:669 [wandb_setup.py:_flush():76] Current SDK version is 0.17.0 +2024-05-23 06:39:44,900 INFO MainThread:669 [wandb_setup.py:_flush():76] Configure stats pid to 669 +2024-05-23 06:39:44,900 INFO MainThread:669 [wandb_setup.py:_flush():76] Loading settings from /root/.config/wandb/settings +2024-05-23 06:39:44,900 INFO MainThread:669 [wandb_setup.py:_flush():76] Loading settings from /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/settings +2024-05-23 06:39:44,900 INFO MainThread:669 [wandb_setup.py:_flush():76] Loading settings from environment variables: {} +2024-05-23 06:39:44,900 INFO MainThread:669 [wandb_setup.py:_flush():76] Applying setup settings: {'_disable_service': False} +2024-05-23 06:39:44,900 WARNING MainThread:669 [wandb_setup.py:_flush():76] Could not find program at -m lm_eval.__main__ +2024-05-23 06:39:44,900 INFO MainThread:669 [wandb_setup.py:_flush():76] Inferring run settings from compute environment: {'program_relpath': None, 'program': '-m lm_eval.__main__'} +2024-05-23 06:39:44,900 INFO MainThread:669 [wandb_setup.py:_flush():76] Applying login settings: {} +2024-05-23 06:39:44,900 INFO MainThread:669 [wandb_init.py:_log_setup():520] Logging user logs to /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240523_063944-3nhz8pu0/logs/debug.log +2024-05-23 06:39:44,900 INFO MainThread:669 [wandb_init.py:_log_setup():521] Logging internal logs to /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240523_063944-3nhz8pu0/logs/debug-internal.log +2024-05-23 06:39:44,900 INFO MainThread:669 [wandb_init.py:init():560] calling init triggers +2024-05-23 06:39:44,900 INFO MainThread:669 [wandb_init.py:init():567] wandb.init called with sweep_config: {} +config: {} +2024-05-23 06:39:44,900 INFO MainThread:669 [wandb_init.py:init():610] starting backend +2024-05-23 06:39:44,900 INFO MainThread:669 [wandb_init.py:init():614] setting up manager +2024-05-23 06:39:44,904 INFO MainThread:669 [backend.py:_multiprocessing_setup():105] multiprocessing start_methods=fork,spawn,forkserver, using: spawn +2024-05-23 06:39:44,906 INFO MainThread:669 [wandb_init.py:init():622] backend started and connected +2024-05-23 06:39:44,909 INFO MainThread:669 [wandb_init.py:init():711] updated telemetry +2024-05-23 06:39:44,918 INFO MainThread:669 [wandb_init.py:init():744] communicating run to backend with 90.0 second timeout +2024-05-23 06:39:45,188 INFO MainThread:669 [wandb_run.py:_on_init():2396] communicating current version +2024-05-23 06:39:45,304 INFO MainThread:669 [wandb_run.py:_on_init():2405] got version response +2024-05-23 06:39:45,304 INFO MainThread:669 [wandb_init.py:init():795] starting run threads in backend +2024-05-23 06:39:45,594 INFO MainThread:669 [wandb_run.py:_console_start():2374] atexit reg +2024-05-23 06:39:45,594 INFO MainThread:669 [wandb_run.py:_redirect():2229] redirect: wrap_raw +2024-05-23 06:39:45,594 INFO MainThread:669 [wandb_run.py:_redirect():2294] Wrapping output streams. +2024-05-23 06:39:45,594 INFO MainThread:669 [wandb_run.py:_redirect():2319] Redirects installed. +2024-05-23 06:39:45,598 INFO MainThread:669 [wandb_init.py:init():838] run started, returning control to user process +2024-05-23 06:40:02,967 WARNING MsgRouterThr:669 [router.py:message_loop():77] message_loop has been closed diff --git a/lm-evaluation-harness/wandb/run-20240523_063944-3nhz8pu0/run-3nhz8pu0.wandb b/lm-evaluation-harness/wandb/run-20240523_063944-3nhz8pu0/run-3nhz8pu0.wandb new file mode 100644 index 0000000000000000000000000000000000000000..d727c0019c713ba5c6188c4d0df5e0dc1c47462b Binary files /dev/null and b/lm-evaluation-harness/wandb/run-20240523_063944-3nhz8pu0/run-3nhz8pu0.wandb differ diff --git a/lm-evaluation-harness/wandb/run-20240523_065043-s0txvu8k/files/config.yaml b/lm-evaluation-harness/wandb/run-20240523_065043-s0txvu8k/files/config.yaml new file mode 100644 index 0000000000000000000000000000000000000000..399b681fec3ea4766ffe7678458635a4e03552bd --- /dev/null +++ b/lm-evaluation-harness/wandb/run-20240523_065043-s0txvu8k/files/config.yaml @@ -0,0 +1,43 @@ +wandb_version: 1 + +_wandb: + desc: null + value: + python_version: 3.10.12 + cli_version: 0.17.0 + framework: huggingface + huggingface_version: 4.41.1 + is_jupyter_run: false + is_kaggle_kernel: false + start_time: 1716447043 + t: + 1: + - 1 + - 5 + - 11 + - 49 + - 51 + - 53 + - 55 + - 71 + - 98 + - 100 + 2: + - 1 + - 5 + - 11 + - 49 + - 51 + - 53 + - 55 + - 71 + - 98 + - 100 + 3: + - 23 + 4: 3.10.12 + 5: 0.17.0 + 6: 4.41.1 + 8: + - 5 + 13: linux-x86_64 diff --git a/lm-evaluation-harness/wandb/run-20240523_065043-s0txvu8k/files/output.log b/lm-evaluation-harness/wandb/run-20240523_065043-s0txvu8k/files/output.log new file mode 100644 index 0000000000000000000000000000000000000000..acbb2f99e0cd814c13ed4c59e96b2232df087c96 --- /dev/null +++ b/lm-evaluation-harness/wandb/run-20240523_065043-s0txvu8k/files/output.log @@ -0,0 +1,34 @@ + +2024-05-23:06:50:43,935 INFO [__main__.py:251] Verbosity set to INFO +2024-05-23:06:50:52,403 INFO [__main__.py:335] Selected Tasks: ['arc_easy', 'hellaswag', 'mrpc', 'openbookqa', 'sst2', 'winogrande'] +2024-05-23:06:50:52,404 INFO [evaluator.py:131] Setting random seed to 0 | Setting numpy seed to 1234 | Setting torch manual seed to 1234 +2024-05-23:06:50:52,404 INFO [evaluator.py:177] Initializing hf model, with arguments: {'pretrained': '/mnt/weka/peacock/experiments/llama/checkpoint/llamav2-3b//hf_ckpt//global_step12000'} +2024-05-23:06:50:54,701 INFO [huggingface.py:164] Using device 'cuda' +Traceback (most recent call last): + File "/usr/lib/python3.10/runpy.py", line 196, in _run_module_as_main + return _run_code(code, main_globals, None, + File "/usr/lib/python3.10/runpy.py", line 86, in _run_code + exec(code, run_globals) + File "/mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/lm_eval/__main__.py", line 417, in + cli_evaluate() + File "/mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/lm_eval/__main__.py", line 341, in cli_evaluate + results = evaluator.simple_evaluate( + File "/mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/lm_eval/utils.py", line 288, in _wrapper + return fn(*args, **kwargs) + File "/mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/lm_eval/evaluator.py", line 180, in simple_evaluate + lm = lm_eval.api.registry.get_model(model).create_from_arg_string( + File "/mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/lm_eval/api/model.py", line 134, in create_from_arg_string + return cls(**args, **args2) + File "/mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/lm_eval/models/huggingface.py", line 190, in __init__ + self._get_config( + File "/mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/lm_eval/models/huggingface.py", line 471, in _get_config + self._config = transformers.AutoConfig.from_pretrained( + File "/usr/local/lib/python3.10/dist-packages/transformers/models/auto/configuration_auto.py", line 934, in from_pretrained + config_dict, unused_kwargs = PretrainedConfig.get_config_dict(pretrained_model_name_or_path, **kwargs) + File "/usr/local/lib/python3.10/dist-packages/transformers/configuration_utils.py", line 632, in get_config_dict + config_dict, kwargs = cls._get_config_dict(pretrained_model_name_or_path, **kwargs) + File "/usr/local/lib/python3.10/dist-packages/transformers/configuration_utils.py", line 689, in _get_config_dict + resolved_config_file = cached_file( + File "/usr/local/lib/python3.10/dist-packages/transformers/utils/hub.py", line 370, in cached_file + raise EnvironmentError( +OSError: /mnt/weka/peacock/experiments/llama/checkpoint/llamav2-3b//hf_ckpt//global_step12000 does not appear to have a file named config.json. Checkout 'https://huggingface.co//mnt/weka/peacock/experiments/llama/checkpoint/llamav2-3b//hf_ckpt//global_step12000/tree/main' for available files. \ No newline at end of file diff --git a/lm-evaluation-harness/wandb/run-20240523_065043-s0txvu8k/files/requirements.txt b/lm-evaluation-harness/wandb/run-20240523_065043-s0txvu8k/files/requirements.txt new file mode 100644 index 0000000000000000000000000000000000000000..094efea889e07cd1194f661d9bac8791e99088d6 --- /dev/null +++ b/lm-evaluation-harness/wandb/run-20240523_065043-s0txvu8k/files/requirements.txt @@ -0,0 +1,154 @@ +DataProperty==1.0.1 +GitPython==3.1.43 +Jinja2==3.1.3 +Markdown==3.6 +MarkupSafe==2.1.5 +Pillow-SIMD==7.0.0.post3 +PyYAML==6.0 +Werkzeug==3.0.2 +absl-py==2.1.0 +accelerate==0.30.1 +aiohttp==3.9.4 +aiosignal==1.3.1 +async-timeout==4.0.3 +attrs==23.2.0 +av==9.2.0 +cachetools==5.3.3 +certifi==2024.2.2 +cffi==1.15.1 +cfgv==3.4.0 +chardet==5.2.0 +charset-normalizer==3.3.2 +click==8.1.7 +cmake==3.29.2 +colorama==0.4.6 +datasets==2.19.1 +deepspeed==0.12.4+hpu.synapse.v1.15.1 +dill==0.3.8 +distlib==0.3.8 +docker-pycreds==0.4.0 +einops==0.8.0 +evaluate==0.4.2 +exceptiongroup==1.2.0 +expecttest==0.2.1 +filelock==3.13.4 +frozenlist==1.4.1 +fsspec==2024.3.1 +gitdb==4.0.11 +google-auth-oauthlib==0.4.6 +google-auth==2.29.0 +grpcio==1.62.1 +habana-media-loader==1.15.1.15 +habana-pyhlml==1.15.1.15 +habana-torch-dataloader==1.15.1.15 +habana-torch-plugin==1.15.1.15 +habana_gpu_migration==1.15.1.15 +habana_quantization_toolkit==1.15.1.15 +hjson==3.1.0 +huggingface-hub==0.23.1 +identify==2.5.35 +idna==3.7 +iniconfig==2.0.0 +joblib==1.4.2 +jsonlines==4.0.0 +lightning-habana==1.4.0 +lightning-utilities==0.11.2 +lightning==2.2.0.post0 +lm_eval==0.4.2 +lm_eval==0.4.2 +lm_eval==0.4.2 +lxml==5.2.2 +mbstrdecoder==1.1.3 +more-itertools==10.2.0 +mpi4py==3.1.4 +mpmath==1.3.0 +multidict==6.0.5 +multiprocess==0.70.16 +networkx==3.3 +ninja==1.11.1.1 +nltk==3.8.1 +nodeenv==1.8.0 +numexpr==2.10.0 +numpy==1.23.5 +oauthlib==3.2.2 +packaging==24.0 +pandas==2.0.1 +pathspec==0.12.1 +pathvalidate==3.2.0 +peft==0.11.1 +perfetto==0.7.0 +pillow==10.3.0 +pip==22.0.2 +pip==23.3.1 +platformdirs==4.2.0 +pluggy==1.4.0 +portalocker==2.8.2 +pre-commit==3.3.3 +protobuf==3.20.3 +psutil==5.9.8 +py-cpuinfo==9.0.0 +pyarrow-hotfix==0.6 +pyarrow==16.1.0 +pyasn1==0.6.0 +pyasn1_modules==0.4.0 +pybind11==2.10.4 +pycparser==2.22 +pydantic==1.10.13 +pynvml==8.0.4 +pytablewriter==1.2.0 +pytest==8.1.1 +python-dateutil==2.9.0.post0 +pytorch-lightning==2.2.2 +pytz==2024.1 +regex==2023.5.5 +requests-oauthlib==2.0.0 +requests==2.31.0 +rouge_score==0.1.2 +rsa==4.9 +sacrebleu==2.4.2 +safetensors==0.4.3 +scikit-learn==1.5.0 +scipy==1.13.1 +sentencepiece==0.2.0 +sentry-sdk==2.2.1 +setproctitle==1.3.3 +setuptools==59.6.0 +setuptools==69.5.1 +six==1.16.0 +smmap==5.0.1 +sqlitedict==2.1.0 +symengine==0.11.0 +sympy==1.12 +tabledata==1.3.3 +tabulate==0.9.0 +tcolorpy==0.1.6 +tdqm==0.0.1 +tensorboard-data-server==0.6.1 +tensorboard-plugin-wit==1.8.1 +tensorboard==2.11.2 +threadpoolctl==3.5.0 +tokenizers==0.19.1 +tomli==2.0.1 +torch==2.2.0a0+git8964477 +torch_tb_profiler==0.4.0 +torchaudio==2.2.0+08901ad +torchdata==0.7.1+5e6f7b7 +torchmetrics==1.3.2 +torchtext==0.17.0+400da5c +torchvision==0.17.0+b2383d4 +tqdm-multiprocess==0.0.11 +tqdm==4.66.2 +transformers==4.41.1 +typepy==1.3.2 +typing_extensions==4.11.0 +tzdata==2024.1 +urllib3==1.26.18 +virtualenv==20.25.1 +wandb==0.17.0 +wheel==0.37.1 +wheel==0.43.0 +word2number==1.1 +xxhash==3.4.1 +yamllint==1.35.1 +yarl==1.9.4 +zstandard==0.22.0 \ No newline at end of file diff --git a/lm-evaluation-harness/wandb/run-20240523_065043-s0txvu8k/files/wandb-metadata.json b/lm-evaluation-harness/wandb/run-20240523_065043-s0txvu8k/files/wandb-metadata.json new file mode 100644 index 0000000000000000000000000000000000000000..0c7889e569b8d1900f194946407976e0c3efc4a8 --- /dev/null +++ b/lm-evaluation-harness/wandb/run-20240523_065043-s0txvu8k/files/wandb-metadata.json @@ -0,0 +1,850 @@ +{ + "os": "Linux-5.15.0-92-generic-x86_64-with-glibc2.35", + "python": "3.10.12", + "heartbeatAt": "2024-05-23T06:50:43.734677", + "startedAt": "2024-05-23T06:50:43.140424", + "docker": null, + "cuda": null, + "args": [ + "--model", + "hf", + "--model_args", + "pretrained=/mnt/weka/peacock/experiments/llama/checkpoint/llamav2-3b//hf_ckpt//global_step12000", + "--tasks", + "hellaswag,arc_easy,openbookqa,winogrande,sst2,mrpc", + "--batch_size", + "auto", + "--wandb_args", + "project=bharatgpt,group=trial_expt_2" + ], + "state": "running", + "program": "-m lm_eval.__main__", + "codePathLocal": null, + "git": { + "remote": "https://github.com/EleutherAI/lm-evaluation-harness", + "commit": null + }, + "email": null, + "root": "/mnt/weka/peacock/idc/cronscript/lm-evaluation-harness", + "host": "peacock-evaluation-worker-0", + "username": "root", + "executable": "/usr/bin/python3", + "cpu_count": 80, + "cpu_count_logical": 160, + "cpu_freq": { + "current": 2326.24500625, + "min": 800.0, + "max": 3400.0 + }, + "cpu_freq_per_core": [ + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3399.997, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + } + ], + "disk": { + "/": { + "total": 877.6341285705566, + "used": 203.88152313232422 + } + }, + "memory": { + "total": 1007.4379196166992 + } +} diff --git a/lm-evaluation-harness/wandb/run-20240523_065043-s0txvu8k/files/wandb-summary.json b/lm-evaluation-harness/wandb/run-20240523_065043-s0txvu8k/files/wandb-summary.json new file mode 100644 index 0000000000000000000000000000000000000000..8bf99d152ad35c3699ec8600ecb8b169d4e35875 --- /dev/null +++ b/lm-evaluation-harness/wandb/run-20240523_065043-s0txvu8k/files/wandb-summary.json @@ -0,0 +1 @@ +{"_wandb": {"runtime": 11}} \ No newline at end of file diff --git a/lm-evaluation-harness/wandb/run-20240523_065043-s0txvu8k/logs/debug-internal.log b/lm-evaluation-harness/wandb/run-20240523_065043-s0txvu8k/logs/debug-internal.log new file mode 100644 index 0000000000000000000000000000000000000000..fb8f3b8a206f6e98068d20cbd1c9a1bcf460864c --- /dev/null +++ b/lm-evaluation-harness/wandb/run-20240523_065043-s0txvu8k/logs/debug-internal.log @@ -0,0 +1,182 @@ +2024-05-23 06:50:43,163 INFO StreamThr :1359 [internal.py:wandb_internal():85] W&B internal server running at pid: 1359, started at: 2024-05-23 06:50:43.162094 +2024-05-23 06:50:43,169 DEBUG HandlerThread:1359 [handler.py:handle_request():158] handle_request: status +2024-05-23 06:50:43,170 INFO WriterThread:1359 [datastore.py:open_for_write():87] open: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240523_065043-s0txvu8k/run-s0txvu8k.wandb +2024-05-23 06:50:43,172 DEBUG SenderThread:1359 [sender.py:send():378] send: header +2024-05-23 06:50:43,175 DEBUG SenderThread:1359 [sender.py:send():378] send: run +2024-05-23 06:50:43,465 INFO SenderThread:1359 [dir_watcher.py:__init__():211] watching files in: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240523_065043-s0txvu8k/files +2024-05-23 06:50:43,465 INFO SenderThread:1359 [sender.py:_start_run_threads():1123] run started: s0txvu8k with start time 1716447043.161806 +2024-05-23 06:50:43,469 DEBUG HandlerThread:1359 [handler.py:handle_request():158] handle_request: check_version +2024-05-23 06:50:43,470 DEBUG SenderThread:1359 [sender.py:send_request():405] send_request: check_version +2024-05-23 06:50:43,658 DEBUG HandlerThread:1359 [handler.py:handle_request():158] handle_request: run_start +2024-05-23 06:50:43,660 DEBUG HandlerThread:1359 [system_info.py:__init__():26] System info init +2024-05-23 06:50:43,660 DEBUG HandlerThread:1359 [system_info.py:__init__():41] System info init done +2024-05-23 06:50:43,660 INFO HandlerThread:1359 [system_monitor.py:start():194] Starting system monitor +2024-05-23 06:50:43,660 INFO SystemMonitor:1359 [system_monitor.py:_start():158] Starting system asset monitoring threads +2024-05-23 06:50:43,660 INFO HandlerThread:1359 [system_monitor.py:probe():214] Collecting system info +2024-05-23 06:50:43,668 INFO SystemMonitor:1359 [interfaces.py:start():188] Started cpu monitoring +2024-05-23 06:50:43,668 INFO SystemMonitor:1359 [interfaces.py:start():188] Started disk monitoring +2024-05-23 06:50:43,669 INFO SystemMonitor:1359 [interfaces.py:start():188] Started memory monitoring +2024-05-23 06:50:43,669 INFO SystemMonitor:1359 [interfaces.py:start():188] Started network monitoring +2024-05-23 06:50:43,734 DEBUG HandlerThread:1359 [system_info.py:probe():150] Probing system +2024-05-23 06:50:43,738 DEBUG HandlerThread:1359 [system_info.py:_probe_git():135] Probing git +2024-05-23 06:50:43,747 ERROR HandlerThread:1359 [gitlib.py:root():92] git root error: Cmd('git') failed due to: exit code(128) + cmdline: git rev-parse --show-toplevel + stderr: 'fatal: detected dubious ownership in repository at '/mnt/weka/peacock/idc/cronscript/lm-evaluation-harness' +To add an exception for this directory, call: + + git config --global --add safe.directory /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness' +2024-05-23 06:50:43,747 DEBUG HandlerThread:1359 [system_info.py:_probe_git():143] Probing git done +2024-05-23 06:50:43,747 DEBUG HandlerThread:1359 [system_info.py:probe():198] Probing system done +2024-05-23 06:50:43,747 DEBUG HandlerThread:1359 [system_monitor.py:probe():223] {'os': 'Linux-5.15.0-92-generic-x86_64-with-glibc2.35', 'python': '3.10.12', 'heartbeatAt': '2024-05-23T06:50:43.734677', 'startedAt': '2024-05-23T06:50:43.140424', 'docker': None, 'cuda': None, 'args': ('--model', 'hf', '--model_args', 'pretrained=/mnt/weka/peacock/experiments/llama/checkpoint/llamav2-3b//hf_ckpt//global_step12000', '--tasks', 'hellaswag,arc_easy,openbookqa,winogrande,sst2,mrpc', '--batch_size', 'auto', '--wandb_args', 'project=bharatgpt,group=trial_expt_2'), 'state': 'running', 'program': '-m lm_eval.__main__', 'codePathLocal': None, 'git': {'remote': 'https://github.com/EleutherAI/lm-evaluation-harness', 'commit': None}, 'email': None, 'root': '/mnt/weka/peacock/idc/cronscript/lm-evaluation-harness', 'host': 'peacock-evaluation-worker-0', 'username': 'root', 'executable': '/usr/bin/python3', 'cpu_count': 80, 'cpu_count_logical': 160, 'cpu_freq': {'current': 2326.24500625, 'min': 800.0, 'max': 3400.0}, 'cpu_freq_per_core': [{'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 3399.997, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}], 'disk': {'/': {'total': 877.6341285705566, 'used': 203.88152313232422}}, 'memory': {'total': 1007.4379196166992}} +2024-05-23 06:50:43,748 INFO HandlerThread:1359 [system_monitor.py:probe():224] Finished collecting system info +2024-05-23 06:50:43,748 INFO HandlerThread:1359 [system_monitor.py:probe():227] Publishing system info +2024-05-23 06:50:43,751 INFO HandlerThread:1359 [system_monitor.py:probe():229] Finished publishing system info +2024-05-23 06:50:43,756 DEBUG SenderThread:1359 [sender.py:send():378] send: files +2024-05-23 06:50:43,756 INFO SenderThread:1359 [sender.py:_save_file():1389] saving file wandb-metadata.json with policy now +2024-05-23 06:50:43,928 DEBUG HandlerThread:1359 [handler.py:handle_request():158] handle_request: python_packages +2024-05-23 06:50:43,929 DEBUG SenderThread:1359 [sender.py:send_request():405] send_request: python_packages +2024-05-23 06:50:43,931 DEBUG SenderThread:1359 [sender.py:send():378] send: telemetry +2024-05-23 06:50:43,931 DEBUG HandlerThread:1359 [handler.py:handle_request():158] handle_request: stop_status +2024-05-23 06:50:43,932 DEBUG SenderThread:1359 [sender.py:send_request():405] send_request: stop_status +2024-05-23 06:50:44,348 INFO wandb-upload_0:1359 [upload_job.py:push():130] Uploaded file /tmp/tmpp9d_n33nwandb/xd4romga-wandb-metadata.json +2024-05-23 06:50:44,467 INFO Thread-12 :1359 [dir_watcher.py:_on_file_created():271] file/dir created: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240523_065043-s0txvu8k/files/output.log +2024-05-23 06:50:44,467 INFO Thread-12 :1359 [dir_watcher.py:_on_file_created():271] file/dir created: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240523_065043-s0txvu8k/files/wandb-metadata.json +2024-05-23 06:50:44,467 INFO Thread-12 :1359 [dir_watcher.py:_on_file_created():271] file/dir created: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240523_065043-s0txvu8k/files/requirements.txt +2024-05-23 06:50:46,467 INFO Thread-12 :1359 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240523_065043-s0txvu8k/files/output.log +2024-05-23 06:50:49,098 DEBUG HandlerThread:1359 [handler.py:handle_request():158] handle_request: status_report +2024-05-23 06:50:54,405 DEBUG HandlerThread:1359 [handler.py:handle_request():158] handle_request: status_report +2024-05-23 06:50:54,476 INFO Thread-12 :1359 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240523_065043-s0txvu8k/files/output.log +2024-05-23 06:50:54,714 DEBUG SenderThread:1359 [sender.py:send():378] send: exit +2024-05-23 06:50:54,714 INFO SenderThread:1359 [sender.py:send_exit():585] handling exit code: 1 +2024-05-23 06:50:54,714 INFO SenderThread:1359 [sender.py:send_exit():587] handling runtime: 11 +2024-05-23 06:50:54,716 INFO SenderThread:1359 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end +2024-05-23 06:50:54,716 INFO SenderThread:1359 [sender.py:send_exit():593] send defer +2024-05-23 06:50:54,716 DEBUG HandlerThread:1359 [handler.py:handle_request():158] handle_request: defer +2024-05-23 06:50:54,716 INFO HandlerThread:1359 [handler.py:handle_request_defer():184] handle defer: 0 +2024-05-23 06:50:54,716 DEBUG SenderThread:1359 [sender.py:send_request():405] send_request: defer +2024-05-23 06:50:54,716 INFO SenderThread:1359 [sender.py:send_request_defer():609] handle sender defer: 0 +2024-05-23 06:50:54,716 INFO SenderThread:1359 [sender.py:transition_state():613] send defer: 1 +2024-05-23 06:50:54,716 DEBUG HandlerThread:1359 [handler.py:handle_request():158] handle_request: defer +2024-05-23 06:50:54,716 INFO HandlerThread:1359 [handler.py:handle_request_defer():184] handle defer: 1 +2024-05-23 06:50:54,717 DEBUG SenderThread:1359 [sender.py:send_request():405] send_request: defer +2024-05-23 06:50:54,717 INFO SenderThread:1359 [sender.py:send_request_defer():609] handle sender defer: 1 +2024-05-23 06:50:54,717 INFO SenderThread:1359 [sender.py:transition_state():613] send defer: 2 +2024-05-23 06:50:54,717 DEBUG HandlerThread:1359 [handler.py:handle_request():158] handle_request: defer +2024-05-23 06:50:54,717 INFO HandlerThread:1359 [handler.py:handle_request_defer():184] handle defer: 2 +2024-05-23 06:50:54,717 INFO HandlerThread:1359 [system_monitor.py:finish():203] Stopping system monitor +2024-05-23 06:50:54,717 DEBUG SystemMonitor:1359 [system_monitor.py:_start():172] Starting system metrics aggregation loop +2024-05-23 06:50:54,717 DEBUG SystemMonitor:1359 [system_monitor.py:_start():179] Finished system metrics aggregation loop +2024-05-23 06:50:54,717 DEBUG SystemMonitor:1359 [system_monitor.py:_start():183] Publishing last batch of metrics +2024-05-23 06:50:54,718 INFO HandlerThread:1359 [interfaces.py:finish():200] Joined cpu monitor +2024-05-23 06:50:54,718 INFO HandlerThread:1359 [interfaces.py:finish():200] Joined disk monitor +2024-05-23 06:50:54,718 INFO HandlerThread:1359 [interfaces.py:finish():200] Joined memory monitor +2024-05-23 06:50:54,718 INFO HandlerThread:1359 [interfaces.py:finish():200] Joined network monitor +2024-05-23 06:50:54,718 DEBUG SenderThread:1359 [sender.py:send_request():405] send_request: defer +2024-05-23 06:50:54,718 INFO SenderThread:1359 [sender.py:send_request_defer():609] handle sender defer: 2 +2024-05-23 06:50:54,718 INFO SenderThread:1359 [sender.py:transition_state():613] send defer: 3 +2024-05-23 06:50:54,718 DEBUG HandlerThread:1359 [handler.py:handle_request():158] handle_request: defer +2024-05-23 06:50:54,718 INFO HandlerThread:1359 [handler.py:handle_request_defer():184] handle defer: 3 +2024-05-23 06:50:54,719 DEBUG SenderThread:1359 [sender.py:send_request():405] send_request: defer +2024-05-23 06:50:54,719 INFO SenderThread:1359 [sender.py:send_request_defer():609] handle sender defer: 3 +2024-05-23 06:50:54,719 INFO SenderThread:1359 [sender.py:transition_state():613] send defer: 4 +2024-05-23 06:50:54,719 DEBUG HandlerThread:1359 [handler.py:handle_request():158] handle_request: defer +2024-05-23 06:50:54,719 INFO HandlerThread:1359 [handler.py:handle_request_defer():184] handle defer: 4 +2024-05-23 06:50:54,719 DEBUG SenderThread:1359 [sender.py:send_request():405] send_request: defer +2024-05-23 06:50:54,719 INFO SenderThread:1359 [sender.py:send_request_defer():609] handle sender defer: 4 +2024-05-23 06:50:54,719 INFO SenderThread:1359 [sender.py:transition_state():613] send defer: 5 +2024-05-23 06:50:54,719 DEBUG HandlerThread:1359 [handler.py:handle_request():158] handle_request: defer +2024-05-23 06:50:54,719 INFO HandlerThread:1359 [handler.py:handle_request_defer():184] handle defer: 5 +2024-05-23 06:50:54,719 DEBUG SenderThread:1359 [sender.py:send():378] send: summary +2024-05-23 06:50:54,720 INFO SenderThread:1359 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end +2024-05-23 06:50:54,720 DEBUG SenderThread:1359 [sender.py:send_request():405] send_request: defer +2024-05-23 06:50:54,720 INFO SenderThread:1359 [sender.py:send_request_defer():609] handle sender defer: 5 +2024-05-23 06:50:54,720 INFO SenderThread:1359 [sender.py:transition_state():613] send defer: 6 +2024-05-23 06:50:54,720 DEBUG HandlerThread:1359 [handler.py:handle_request():158] handle_request: defer +2024-05-23 06:50:54,720 INFO HandlerThread:1359 [handler.py:handle_request_defer():184] handle defer: 6 +2024-05-23 06:50:54,720 DEBUG SenderThread:1359 [sender.py:send_request():405] send_request: defer +2024-05-23 06:50:54,721 INFO SenderThread:1359 [sender.py:send_request_defer():609] handle sender defer: 6 +2024-05-23 06:50:54,725 DEBUG HandlerThread:1359 [handler.py:handle_request():158] handle_request: status_report +2024-05-23 06:50:54,792 INFO SenderThread:1359 [sender.py:transition_state():613] send defer: 7 +2024-05-23 06:50:54,792 DEBUG HandlerThread:1359 [handler.py:handle_request():158] handle_request: defer +2024-05-23 06:50:54,792 INFO HandlerThread:1359 [handler.py:handle_request_defer():184] handle defer: 7 +2024-05-23 06:50:54,792 DEBUG SenderThread:1359 [sender.py:send_request():405] send_request: defer +2024-05-23 06:50:54,793 INFO SenderThread:1359 [sender.py:send_request_defer():609] handle sender defer: 7 +2024-05-23 06:50:55,478 INFO Thread-12 :1359 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240523_065043-s0txvu8k/files/config.yaml +2024-05-23 06:50:55,478 INFO Thread-12 :1359 [dir_watcher.py:_on_file_created():271] file/dir created: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240523_065043-s0txvu8k/files/wandb-summary.json +2024-05-23 06:50:55,714 DEBUG HandlerThread:1359 [handler.py:handle_request():158] handle_request: poll_exit +2024-05-23 06:50:56,118 INFO SenderThread:1359 [sender.py:transition_state():613] send defer: 8 +2024-05-23 06:50:56,118 DEBUG SenderThread:1359 [sender.py:send_request():405] send_request: poll_exit +2024-05-23 06:50:56,118 DEBUG HandlerThread:1359 [handler.py:handle_request():158] handle_request: defer +2024-05-23 06:50:56,119 INFO HandlerThread:1359 [handler.py:handle_request_defer():184] handle defer: 8 +2024-05-23 06:50:56,119 DEBUG SenderThread:1359 [sender.py:send_request():405] send_request: defer +2024-05-23 06:50:56,119 INFO SenderThread:1359 [sender.py:send_request_defer():609] handle sender defer: 8 +2024-05-23 06:50:56,119 INFO SenderThread:1359 [job_builder.py:build():432] Attempting to build job artifact +2024-05-23 06:50:56,120 INFO SenderThread:1359 [job_builder.py:_get_source_type():576] no source found +2024-05-23 06:50:56,120 INFO SenderThread:1359 [sender.py:transition_state():613] send defer: 9 +2024-05-23 06:50:56,120 DEBUG HandlerThread:1359 [handler.py:handle_request():158] handle_request: defer +2024-05-23 06:50:56,120 INFO HandlerThread:1359 [handler.py:handle_request_defer():184] handle defer: 9 +2024-05-23 06:50:56,120 DEBUG SenderThread:1359 [sender.py:send_request():405] send_request: defer +2024-05-23 06:50:56,120 INFO SenderThread:1359 [sender.py:send_request_defer():609] handle sender defer: 9 +2024-05-23 06:50:56,120 INFO SenderThread:1359 [dir_watcher.py:finish():358] shutting down directory watcher +2024-05-23 06:50:56,479 INFO SenderThread:1359 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240523_065043-s0txvu8k/files/output.log +2024-05-23 06:50:56,480 INFO SenderThread:1359 [dir_watcher.py:finish():388] scan: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240523_065043-s0txvu8k/files +2024-05-23 06:50:56,480 INFO SenderThread:1359 [dir_watcher.py:finish():402] scan save: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240523_065043-s0txvu8k/files/wandb-metadata.json wandb-metadata.json +2024-05-23 06:50:56,480 INFO SenderThread:1359 [dir_watcher.py:finish():402] scan save: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240523_065043-s0txvu8k/files/config.yaml config.yaml +2024-05-23 06:50:56,480 INFO SenderThread:1359 [dir_watcher.py:finish():402] scan save: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240523_065043-s0txvu8k/files/output.log output.log +2024-05-23 06:50:56,482 INFO SenderThread:1359 [dir_watcher.py:finish():402] scan save: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240523_065043-s0txvu8k/files/wandb-summary.json wandb-summary.json +2024-05-23 06:50:56,483 INFO SenderThread:1359 [dir_watcher.py:finish():402] scan save: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240523_065043-s0txvu8k/files/requirements.txt requirements.txt +2024-05-23 06:50:56,483 INFO SenderThread:1359 [sender.py:transition_state():613] send defer: 10 +2024-05-23 06:50:56,483 DEBUG HandlerThread:1359 [handler.py:handle_request():158] handle_request: defer +2024-05-23 06:50:56,483 INFO HandlerThread:1359 [handler.py:handle_request_defer():184] handle defer: 10 +2024-05-23 06:50:56,485 DEBUG SenderThread:1359 [sender.py:send_request():405] send_request: defer +2024-05-23 06:50:56,485 INFO SenderThread:1359 [sender.py:send_request_defer():609] handle sender defer: 10 +2024-05-23 06:50:56,485 INFO SenderThread:1359 [file_pusher.py:finish():169] shutting down file pusher +2024-05-23 06:50:56,714 DEBUG HandlerThread:1359 [handler.py:handle_request():158] handle_request: poll_exit +2024-05-23 06:50:56,714 DEBUG SenderThread:1359 [sender.py:send_request():405] send_request: poll_exit +2024-05-23 06:50:56,717 INFO wandb-upload_0:1359 [upload_job.py:push():130] Uploaded file /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240523_065043-s0txvu8k/files/config.yaml +2024-05-23 06:50:57,086 INFO wandb-upload_3:1359 [upload_job.py:push():130] Uploaded file /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240523_065043-s0txvu8k/files/requirements.txt +2024-05-23 06:50:57,091 INFO wandb-upload_2:1359 [upload_job.py:push():130] Uploaded file /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240523_065043-s0txvu8k/files/wandb-summary.json +2024-05-23 06:50:57,094 INFO wandb-upload_1:1359 [upload_job.py:push():130] Uploaded file /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240523_065043-s0txvu8k/files/output.log +2024-05-23 06:50:57,294 INFO Thread-11 (_thread_body):1359 [sender.py:transition_state():613] send defer: 11 +2024-05-23 06:50:57,295 DEBUG HandlerThread:1359 [handler.py:handle_request():158] handle_request: defer +2024-05-23 06:50:57,295 INFO HandlerThread:1359 [handler.py:handle_request_defer():184] handle defer: 11 +2024-05-23 06:50:57,295 DEBUG SenderThread:1359 [sender.py:send_request():405] send_request: defer +2024-05-23 06:50:57,295 INFO SenderThread:1359 [sender.py:send_request_defer():609] handle sender defer: 11 +2024-05-23 06:50:57,295 INFO SenderThread:1359 [file_pusher.py:join():175] waiting for file pusher +2024-05-23 06:50:57,295 INFO SenderThread:1359 [sender.py:transition_state():613] send defer: 12 +2024-05-23 06:50:57,295 DEBUG HandlerThread:1359 [handler.py:handle_request():158] handle_request: defer +2024-05-23 06:50:57,295 INFO HandlerThread:1359 [handler.py:handle_request_defer():184] handle defer: 12 +2024-05-23 06:50:57,295 DEBUG SenderThread:1359 [sender.py:send_request():405] send_request: defer +2024-05-23 06:50:57,295 INFO SenderThread:1359 [sender.py:send_request_defer():609] handle sender defer: 12 +2024-05-23 06:50:57,295 INFO SenderThread:1359 [file_stream.py:finish():601] file stream finish called +2024-05-23 06:50:57,354 INFO SenderThread:1359 [file_stream.py:finish():605] file stream finish is done +2024-05-23 06:50:57,354 INFO SenderThread:1359 [sender.py:transition_state():613] send defer: 13 +2024-05-23 06:50:57,354 DEBUG HandlerThread:1359 [handler.py:handle_request():158] handle_request: defer +2024-05-23 06:50:57,354 INFO HandlerThread:1359 [handler.py:handle_request_defer():184] handle defer: 13 +2024-05-23 06:50:57,354 DEBUG SenderThread:1359 [sender.py:send_request():405] send_request: defer +2024-05-23 06:50:57,355 INFO SenderThread:1359 [sender.py:send_request_defer():609] handle sender defer: 13 +2024-05-23 06:50:57,355 INFO SenderThread:1359 [sender.py:transition_state():613] send defer: 14 +2024-05-23 06:50:57,355 DEBUG HandlerThread:1359 [handler.py:handle_request():158] handle_request: defer +2024-05-23 06:50:57,355 INFO HandlerThread:1359 [handler.py:handle_request_defer():184] handle defer: 14 +2024-05-23 06:50:57,355 DEBUG SenderThread:1359 [sender.py:send():378] send: final +2024-05-23 06:50:57,355 DEBUG SenderThread:1359 [sender.py:send():378] send: footer +2024-05-23 06:50:57,355 DEBUG SenderThread:1359 [sender.py:send_request():405] send_request: defer +2024-05-23 06:50:57,355 INFO SenderThread:1359 [sender.py:send_request_defer():609] handle sender defer: 14 +2024-05-23 06:50:57,356 DEBUG HandlerThread:1359 [handler.py:handle_request():158] handle_request: poll_exit +2024-05-23 06:50:57,356 DEBUG HandlerThread:1359 [handler.py:handle_request():158] handle_request: poll_exit +2024-05-23 06:50:57,356 DEBUG HandlerThread:1359 [handler.py:handle_request():158] handle_request: server_info +2024-05-23 06:50:57,356 DEBUG HandlerThread:1359 [handler.py:handle_request():158] handle_request: get_summary +2024-05-23 06:50:57,356 DEBUG HandlerThread:1359 [handler.py:handle_request():158] handle_request: sampled_history +2024-05-23 06:50:57,356 DEBUG HandlerThread:1359 [handler.py:handle_request():158] handle_request: internal_messages +2024-05-23 06:50:57,356 DEBUG SenderThread:1359 [sender.py:send_request():405] send_request: poll_exit +2024-05-23 06:50:57,357 DEBUG SenderThread:1359 [sender.py:send_request():405] send_request: poll_exit +2024-05-23 06:50:57,357 DEBUG SenderThread:1359 [sender.py:send_request():405] send_request: server_info +2024-05-23 06:50:57,409 INFO MainThread:1359 [wandb_run.py:_footer_history_summary_info():3994] rendering history +2024-05-23 06:50:57,409 INFO MainThread:1359 [wandb_run.py:_footer_history_summary_info():4026] rendering summary +2024-05-23 06:50:57,409 INFO MainThread:1359 [wandb_run.py:_footer_sync_info():3953] logging synced files +2024-05-23 06:50:57,410 DEBUG HandlerThread:1359 [handler.py:handle_request():158] handle_request: shutdown +2024-05-23 06:50:57,410 INFO HandlerThread:1359 [handler.py:finish():882] shutting down handler +2024-05-23 06:50:58,356 INFO WriterThread:1359 [datastore.py:close():296] close: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240523_065043-s0txvu8k/run-s0txvu8k.wandb +2024-05-23 06:50:58,409 INFO SenderThread:1359 [sender.py:finish():1545] shutting down sender +2024-05-23 06:50:58,409 INFO SenderThread:1359 [file_pusher.py:finish():169] shutting down file pusher +2024-05-23 06:50:58,409 INFO SenderThread:1359 [file_pusher.py:join():175] waiting for file pusher diff --git a/lm-evaluation-harness/wandb/run-20240523_065043-s0txvu8k/logs/debug.log b/lm-evaluation-harness/wandb/run-20240523_065043-s0txvu8k/logs/debug.log new file mode 100644 index 0000000000000000000000000000000000000000..2666a768127cc46a890de3072c8dcfe7c94222e0 --- /dev/null +++ b/lm-evaluation-harness/wandb/run-20240523_065043-s0txvu8k/logs/debug.log @@ -0,0 +1,29 @@ +2024-05-23 06:50:43,156 INFO MainThread:1204 [wandb_setup.py:_flush():76] Current SDK version is 0.17.0 +2024-05-23 06:50:43,156 INFO MainThread:1204 [wandb_setup.py:_flush():76] Configure stats pid to 1204 +2024-05-23 06:50:43,156 INFO MainThread:1204 [wandb_setup.py:_flush():76] Loading settings from /root/.config/wandb/settings +2024-05-23 06:50:43,156 INFO MainThread:1204 [wandb_setup.py:_flush():76] Loading settings from /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/settings +2024-05-23 06:50:43,156 INFO MainThread:1204 [wandb_setup.py:_flush():76] Loading settings from environment variables: {} +2024-05-23 06:50:43,156 INFO MainThread:1204 [wandb_setup.py:_flush():76] Applying setup settings: {'_disable_service': False} +2024-05-23 06:50:43,156 WARNING MainThread:1204 [wandb_setup.py:_flush():76] Could not find program at -m lm_eval.__main__ +2024-05-23 06:50:43,156 INFO MainThread:1204 [wandb_setup.py:_flush():76] Inferring run settings from compute environment: {'program_relpath': None, 'program': '-m lm_eval.__main__'} +2024-05-23 06:50:43,156 INFO MainThread:1204 [wandb_setup.py:_flush():76] Applying login settings: {} +2024-05-23 06:50:43,156 INFO MainThread:1204 [wandb_init.py:_log_setup():520] Logging user logs to /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240523_065043-s0txvu8k/logs/debug.log +2024-05-23 06:50:43,156 INFO MainThread:1204 [wandb_init.py:_log_setup():521] Logging internal logs to /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240523_065043-s0txvu8k/logs/debug-internal.log +2024-05-23 06:50:43,156 INFO MainThread:1204 [wandb_init.py:init():560] calling init triggers +2024-05-23 06:50:43,156 INFO MainThread:1204 [wandb_init.py:init():567] wandb.init called with sweep_config: {} +config: {} +2024-05-23 06:50:43,156 INFO MainThread:1204 [wandb_init.py:init():610] starting backend +2024-05-23 06:50:43,156 INFO MainThread:1204 [wandb_init.py:init():614] setting up manager +2024-05-23 06:50:43,159 INFO MainThread:1204 [backend.py:_multiprocessing_setup():105] multiprocessing start_methods=fork,spawn,forkserver, using: spawn +2024-05-23 06:50:43,161 INFO MainThread:1204 [wandb_init.py:init():622] backend started and connected +2024-05-23 06:50:43,165 INFO MainThread:1204 [wandb_init.py:init():711] updated telemetry +2024-05-23 06:50:43,174 INFO MainThread:1204 [wandb_init.py:init():744] communicating run to backend with 90.0 second timeout +2024-05-23 06:50:43,469 INFO MainThread:1204 [wandb_run.py:_on_init():2396] communicating current version +2024-05-23 06:50:43,651 INFO MainThread:1204 [wandb_run.py:_on_init():2405] got version response +2024-05-23 06:50:43,652 INFO MainThread:1204 [wandb_init.py:init():795] starting run threads in backend +2024-05-23 06:50:43,929 INFO MainThread:1204 [wandb_run.py:_console_start():2374] atexit reg +2024-05-23 06:50:43,930 INFO MainThread:1204 [wandb_run.py:_redirect():2229] redirect: wrap_raw +2024-05-23 06:50:43,930 INFO MainThread:1204 [wandb_run.py:_redirect():2294] Wrapping output streams. +2024-05-23 06:50:43,930 INFO MainThread:1204 [wandb_run.py:_redirect():2319] Redirects installed. +2024-05-23 06:50:43,933 INFO MainThread:1204 [wandb_init.py:init():838] run started, returning control to user process +2024-05-23 06:50:58,411 WARNING MsgRouterThr:1204 [router.py:message_loop():77] message_loop has been closed diff --git a/lm-evaluation-harness/wandb/run-20240523_065043-s0txvu8k/run-s0txvu8k.wandb b/lm-evaluation-harness/wandb/run-20240523_065043-s0txvu8k/run-s0txvu8k.wandb new file mode 100644 index 0000000000000000000000000000000000000000..784be0ef69fd801a8fdecab07de8379c467b0d0d Binary files /dev/null and b/lm-evaluation-harness/wandb/run-20240523_065043-s0txvu8k/run-s0txvu8k.wandb differ diff --git a/lm-evaluation-harness/wandb/run-20240523_123151-6w9nn173/files/config.yaml b/lm-evaluation-harness/wandb/run-20240523_123151-6w9nn173/files/config.yaml new file mode 100644 index 0000000000000000000000000000000000000000..cab2283fe7c98fa985500b225c71a9a77dc13bec --- /dev/null +++ b/lm-evaluation-harness/wandb/run-20240523_123151-6w9nn173/files/config.yaml @@ -0,0 +1,43 @@ +wandb_version: 1 + +_wandb: + desc: null + value: + python_version: 3.10.12 + cli_version: 0.17.0 + framework: huggingface + huggingface_version: 4.41.1 + is_jupyter_run: false + is_kaggle_kernel: false + start_time: 1716467511 + t: + 1: + - 1 + - 5 + - 11 + - 49 + - 51 + - 53 + - 55 + - 71 + - 98 + - 100 + 2: + - 1 + - 5 + - 11 + - 49 + - 51 + - 53 + - 55 + - 71 + - 98 + - 100 + 3: + - 23 + 4: 3.10.12 + 5: 0.17.0 + 6: 4.41.1 + 8: + - 5 + 13: linux-x86_64 diff --git a/lm-evaluation-harness/wandb/run-20240523_123151-6w9nn173/files/output.log b/lm-evaluation-harness/wandb/run-20240523_123151-6w9nn173/files/output.log new file mode 100644 index 0000000000000000000000000000000000000000..7f58bab83e441daf13a59772c0db92c4e6b9d454 --- /dev/null +++ b/lm-evaluation-harness/wandb/run-20240523_123151-6w9nn173/files/output.log @@ -0,0 +1,34 @@ + +2024-05-23:12:31:51,966 INFO [__main__.py:251] Verbosity set to INFO +2024-05-23:12:32:00,580 INFO [__main__.py:335] Selected Tasks: ['arc_easy', 'hellaswag', 'mrpc', 'openbookqa', 'sst2', 'winogrande'] +2024-05-23:12:32:00,581 INFO [evaluator.py:131] Setting random seed to 0 | Setting numpy seed to 1234 | Setting torch manual seed to 1234 +2024-05-23:12:32:00,581 INFO [evaluator.py:177] Initializing hf model, with arguments: {'pretrained': '/mnt/weka/peacock/experiments/llama/checkpoint/llamav2-3b//hf_ckpt//global_step16000'} +2024-05-23:12:32:02,885 INFO [huggingface.py:164] Using device 'cuda' +Traceback (most recent call last): + File "/usr/lib/python3.10/runpy.py", line 196, in _run_module_as_main + return _run_code(code, main_globals, None, + File "/usr/lib/python3.10/runpy.py", line 86, in _run_code + exec(code, run_globals) + File "/mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/lm_eval/__main__.py", line 417, in + cli_evaluate() + File "/mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/lm_eval/__main__.py", line 341, in cli_evaluate + results = evaluator.simple_evaluate( + File "/mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/lm_eval/utils.py", line 288, in _wrapper + return fn(*args, **kwargs) + File "/mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/lm_eval/evaluator.py", line 180, in simple_evaluate + lm = lm_eval.api.registry.get_model(model).create_from_arg_string( + File "/mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/lm_eval/api/model.py", line 134, in create_from_arg_string + return cls(**args, **args2) + File "/mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/lm_eval/models/huggingface.py", line 190, in __init__ + self._get_config( + File "/mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/lm_eval/models/huggingface.py", line 471, in _get_config + self._config = transformers.AutoConfig.from_pretrained( + File "/usr/local/lib/python3.10/dist-packages/transformers/models/auto/configuration_auto.py", line 934, in from_pretrained + config_dict, unused_kwargs = PretrainedConfig.get_config_dict(pretrained_model_name_or_path, **kwargs) + File "/usr/local/lib/python3.10/dist-packages/transformers/configuration_utils.py", line 632, in get_config_dict + config_dict, kwargs = cls._get_config_dict(pretrained_model_name_or_path, **kwargs) + File "/usr/local/lib/python3.10/dist-packages/transformers/configuration_utils.py", line 689, in _get_config_dict + resolved_config_file = cached_file( + File "/usr/local/lib/python3.10/dist-packages/transformers/utils/hub.py", line 370, in cached_file + raise EnvironmentError( +OSError: /mnt/weka/peacock/experiments/llama/checkpoint/llamav2-3b//hf_ckpt//global_step16000 does not appear to have a file named config.json. Checkout 'https://huggingface.co//mnt/weka/peacock/experiments/llama/checkpoint/llamav2-3b//hf_ckpt//global_step16000/tree/main' for available files. \ No newline at end of file diff --git a/lm-evaluation-harness/wandb/run-20240523_123151-6w9nn173/files/requirements.txt b/lm-evaluation-harness/wandb/run-20240523_123151-6w9nn173/files/requirements.txt new file mode 100644 index 0000000000000000000000000000000000000000..f675c3016b5332c1acf28f436e0b60adeead9c12 --- /dev/null +++ b/lm-evaluation-harness/wandb/run-20240523_123151-6w9nn173/files/requirements.txt @@ -0,0 +1,155 @@ +DataProperty==1.0.1 +GitPython==3.1.43 +Jinja2==3.1.4 +Markdown==3.6 +MarkupSafe==2.1.5 +Pillow-SIMD==7.0.0.post3 +PyYAML==6.0 +Werkzeug==3.0.3 +absl-py==2.1.0 +accelerate==0.30.1 +aiohttp==3.9.5 +aiosignal==1.3.1 +async-timeout==4.0.3 +attrs==23.2.0 +av==9.2.0 +cachetools==5.3.3 +certifi==2024.2.2 +cffi==1.15.1 +cfgv==3.4.0 +chardet==5.2.0 +charset-normalizer==3.3.2 +click==8.1.7 +cmake==3.29.2 +colorama==0.4.6 +datasets==2.19.1 +deepspeed==0.12.4+hpu.synapse.v1.15.1 +dill==0.3.8 +distlib==0.3.8 +docker-pycreds==0.4.0 +einops==0.8.0 +evaluate==0.4.2 +exceptiongroup==1.2.1 +expecttest==0.2.1 +filelock==3.14.0 +frozenlist==1.4.1 +fsspec==2024.3.1 +gitdb==4.0.11 +google-auth-oauthlib==0.4.6 +google-auth==2.29.0 +grpcio==1.63.0 +habana-media-loader==1.15.1.15 +habana-pyhlml==1.15.1.15 +habana-torch-dataloader==1.15.1.15 +habana-torch-plugin==1.15.1.15 +habana_gpu_migration==1.15.1.15 +habana_quantization_toolkit==1.15.1.15 +hjson==3.1.0 +huggingface-hub==0.23.1 +identify==2.5.36 +idna==3.7 +iniconfig==2.0.0 +joblib==1.4.2 +jsonlines==4.0.0 +lightning-habana==1.4.0 +lightning-utilities==0.11.2 +lightning==2.2.0.post0 +lm_eval==0.4.2 +lm_eval==0.4.2 +lm_eval==0.4.2 +lxml==5.2.2 +mbstrdecoder==1.1.3 +more-itertools==10.2.0 +mpi4py==3.1.4 +mpmath==1.3.0 +multidict==6.0.5 +multiprocess==0.70.16 +networkx==3.3 +ninja==1.11.1.1 +nltk==3.8.1 +nodeenv==1.8.0 +numexpr==2.10.0 +numpy==1.23.5 +oauthlib==3.2.2 +packaging==24.0 +pandas==2.0.1 +pathspec==0.12.1 +pathvalidate==3.2.0 +peft==0.11.1 +perfetto==0.7.0 +pillow==10.3.0 +pip==22.0.2 +pip==23.3.1 +platformdirs==4.2.1 +pluggy==1.5.0 +portalocker==2.8.2 +pre-commit==3.3.3 +pretty-errors==1.2.25 +protobuf==3.20.3 +psutil==5.9.8 +py-cpuinfo==9.0.0 +pyarrow-hotfix==0.6 +pyarrow==16.1.0 +pyasn1==0.6.0 +pyasn1_modules==0.4.0 +pybind11==2.10.4 +pycparser==2.22 +pydantic==1.10.13 +pynvml==8.0.4 +pytablewriter==1.2.0 +pytest==8.2.0 +python-dateutil==2.9.0.post0 +pytorch-lightning==2.2.4 +pytz==2024.1 +regex==2023.5.5 +requests-oauthlib==2.0.0 +requests==2.31.0 +rouge_score==0.1.2 +rsa==4.9 +sacrebleu==2.4.2 +safetensors==0.4.3 +scikit-learn==1.5.0 +scipy==1.13.1 +sentencepiece==0.2.0 +sentry-sdk==2.3.0 +setproctitle==1.3.3 +setuptools==59.6.0 +setuptools==69.5.1 +six==1.16.0 +smmap==5.0.1 +sqlitedict==2.1.0 +symengine==0.11.0 +sympy==1.12 +tabledata==1.3.3 +tabulate==0.9.0 +tcolorpy==0.1.6 +tdqm==0.0.1 +tensorboard-data-server==0.6.1 +tensorboard-plugin-wit==1.8.1 +tensorboard==2.11.2 +threadpoolctl==3.5.0 +tokenizers==0.19.1 +tomli==2.0.1 +torch==2.2.0a0+git8964477 +torch_tb_profiler==0.4.0 +torchaudio==2.2.0+08901ad +torchdata==0.7.1+5e6f7b7 +torchmetrics==1.4.0 +torchtext==0.17.0+400da5c +torchvision==0.17.0+b2383d4 +tqdm-multiprocess==0.0.11 +tqdm==4.66.4 +transformers==4.41.1 +typepy==1.3.2 +typing_extensions==4.11.0 +tzdata==2024.1 +urllib3==1.26.18 +virtualenv==20.26.1 +wandb==0.17.0 +wheel==0.37.1 +wheel==0.43.0 +word2number==1.1 +xxhash==3.4.1 +yamllint==1.35.1 +yarl==1.9.4 +zstandard==0.22.0 \ No newline at end of file diff --git a/lm-evaluation-harness/wandb/run-20240523_123151-6w9nn173/files/wandb-metadata.json b/lm-evaluation-harness/wandb/run-20240523_123151-6w9nn173/files/wandb-metadata.json new file mode 100644 index 0000000000000000000000000000000000000000..4718075e215b6c5f26d94641ada74c7d3501755f --- /dev/null +++ b/lm-evaluation-harness/wandb/run-20240523_123151-6w9nn173/files/wandb-metadata.json @@ -0,0 +1,850 @@ +{ + "os": "Linux-5.15.0-92-generic-x86_64-with-glibc2.35", + "python": "3.10.12", + "heartbeatAt": "2024-05-23T12:31:51.757913", + "startedAt": "2024-05-23T12:31:51.159684", + "docker": null, + "cuda": null, + "args": [ + "--model", + "hf", + "--model_args", + "pretrained=/mnt/weka/peacock/experiments/llama/checkpoint/llamav2-3b//hf_ckpt//global_step16000", + "--tasks", + "hellaswag,arc_easy,openbookqa,winogrande,sst2,mrpc", + "--batch_size", + "auto", + "--wandb_args", + "project=bharatgpt,group=trial_expt_2" + ], + "state": "running", + "program": "-m lm_eval.__main__", + "codePathLocal": null, + "git": { + "remote": "https://github.com/EleutherAI/lm-evaluation-harness", + "commit": null + }, + "email": null, + "root": "/mnt/weka/peacock/idc/cronscript/lm-evaluation-harness", + "host": "peacock-evaluation-worker-0", + "username": "root", + "executable": "/usr/bin/python3", + "cpu_count": 80, + "cpu_count_logical": 160, + "cpu_freq": { + "current": 2334.2906312500004, + "min": 800.0, + "max": 3400.0 + }, + "cpu_freq_per_core": [ + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + } + ], + "disk": { + "/": { + "total": 877.6341285705566, + "used": 209.58220291137695 + } + }, + "memory": { + "total": 1007.4379425048828 + } +} diff --git a/lm-evaluation-harness/wandb/run-20240523_123151-6w9nn173/files/wandb-summary.json b/lm-evaluation-harness/wandb/run-20240523_123151-6w9nn173/files/wandb-summary.json new file mode 100644 index 0000000000000000000000000000000000000000..8bf99d152ad35c3699ec8600ecb8b169d4e35875 --- /dev/null +++ b/lm-evaluation-harness/wandb/run-20240523_123151-6w9nn173/files/wandb-summary.json @@ -0,0 +1 @@ +{"_wandb": {"runtime": 11}} \ No newline at end of file diff --git a/lm-evaluation-harness/wandb/run-20240523_123151-6w9nn173/logs/debug-internal.log b/lm-evaluation-harness/wandb/run-20240523_123151-6w9nn173/logs/debug-internal.log new file mode 100644 index 0000000000000000000000000000000000000000..5bba4edca305a7203b31a7e95ddb47eb87b992f8 --- /dev/null +++ b/lm-evaluation-harness/wandb/run-20240523_123151-6w9nn173/logs/debug-internal.log @@ -0,0 +1,182 @@ +2024-05-23 12:31:51,180 INFO StreamThr :1833 [internal.py:wandb_internal():85] W&B internal server running at pid: 1833, started at: 2024-05-23 12:31:51.178170 +2024-05-23 12:31:51,185 DEBUG HandlerThread:1833 [handler.py:handle_request():158] handle_request: status +2024-05-23 12:31:51,187 INFO WriterThread:1833 [datastore.py:open_for_write():87] open: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240523_123151-6w9nn173/run-6w9nn173.wandb +2024-05-23 12:31:51,190 DEBUG SenderThread:1833 [sender.py:send():378] send: header +2024-05-23 12:31:51,190 DEBUG SenderThread:1833 [sender.py:send():378] send: run +2024-05-23 12:31:51,541 INFO SenderThread:1833 [dir_watcher.py:__init__():211] watching files in: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240523_123151-6w9nn173/files +2024-05-23 12:31:51,541 INFO SenderThread:1833 [sender.py:_start_run_threads():1123] run started: 6w9nn173 with start time 1716467511.178643 +2024-05-23 12:31:51,545 DEBUG HandlerThread:1833 [handler.py:handle_request():158] handle_request: check_version +2024-05-23 12:31:51,546 DEBUG SenderThread:1833 [sender.py:send_request():405] send_request: check_version +2024-05-23 12:31:51,662 DEBUG HandlerThread:1833 [handler.py:handle_request():158] handle_request: run_start +2024-05-23 12:31:51,665 DEBUG HandlerThread:1833 [system_info.py:__init__():26] System info init +2024-05-23 12:31:51,665 DEBUG HandlerThread:1833 [system_info.py:__init__():41] System info init done +2024-05-23 12:31:51,665 INFO HandlerThread:1833 [system_monitor.py:start():194] Starting system monitor +2024-05-23 12:31:51,665 INFO SystemMonitor:1833 [system_monitor.py:_start():158] Starting system asset monitoring threads +2024-05-23 12:31:51,665 INFO HandlerThread:1833 [system_monitor.py:probe():214] Collecting system info +2024-05-23 12:31:51,672 INFO SystemMonitor:1833 [interfaces.py:start():188] Started cpu monitoring +2024-05-23 12:31:51,672 INFO SystemMonitor:1833 [interfaces.py:start():188] Started disk monitoring +2024-05-23 12:31:51,672 INFO SystemMonitor:1833 [interfaces.py:start():188] Started memory monitoring +2024-05-23 12:31:51,673 INFO SystemMonitor:1833 [interfaces.py:start():188] Started network monitoring +2024-05-23 12:31:51,757 DEBUG HandlerThread:1833 [system_info.py:probe():150] Probing system +2024-05-23 12:31:51,760 DEBUG HandlerThread:1833 [system_info.py:_probe_git():135] Probing git +2024-05-23 12:31:51,770 ERROR HandlerThread:1833 [gitlib.py:root():92] git root error: Cmd('git') failed due to: exit code(128) + cmdline: git rev-parse --show-toplevel + stderr: 'fatal: detected dubious ownership in repository at '/mnt/weka/peacock/idc/cronscript/lm-evaluation-harness' +To add an exception for this directory, call: + + git config --global --add safe.directory /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness' +2024-05-23 12:31:51,770 DEBUG HandlerThread:1833 [system_info.py:_probe_git():143] Probing git done +2024-05-23 12:31:51,770 DEBUG HandlerThread:1833 [system_info.py:probe():198] Probing system done +2024-05-23 12:31:51,770 DEBUG HandlerThread:1833 [system_monitor.py:probe():223] {'os': 'Linux-5.15.0-92-generic-x86_64-with-glibc2.35', 'python': '3.10.12', 'heartbeatAt': '2024-05-23T12:31:51.757913', 'startedAt': '2024-05-23T12:31:51.159684', 'docker': None, 'cuda': None, 'args': ('--model', 'hf', '--model_args', 'pretrained=/mnt/weka/peacock/experiments/llama/checkpoint/llamav2-3b//hf_ckpt//global_step16000', '--tasks', 'hellaswag,arc_easy,openbookqa,winogrande,sst2,mrpc', '--batch_size', 'auto', '--wandb_args', 'project=bharatgpt,group=trial_expt_2'), 'state': 'running', 'program': '-m lm_eval.__main__', 'codePathLocal': None, 'git': {'remote': 'https://github.com/EleutherAI/lm-evaluation-harness', 'commit': None}, 'email': None, 'root': '/mnt/weka/peacock/idc/cronscript/lm-evaluation-harness', 'host': 'peacock-evaluation-worker-0', 'username': 'root', 'executable': '/usr/bin/python3', 'cpu_count': 80, 'cpu_count_logical': 160, 'cpu_freq': {'current': 2334.2906312500004, 'min': 800.0, 'max': 3400.0}, 'cpu_freq_per_core': [{'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}], 'disk': {'/': {'total': 877.6341285705566, 'used': 209.58220291137695}}, 'memory': {'total': 1007.4379425048828}} +2024-05-23 12:31:51,770 INFO HandlerThread:1833 [system_monitor.py:probe():224] Finished collecting system info +2024-05-23 12:31:51,770 INFO HandlerThread:1833 [system_monitor.py:probe():227] Publishing system info +2024-05-23 12:31:51,773 INFO HandlerThread:1833 [system_monitor.py:probe():229] Finished publishing system info +2024-05-23 12:31:51,779 DEBUG SenderThread:1833 [sender.py:send():378] send: files +2024-05-23 12:31:51,779 INFO SenderThread:1833 [sender.py:_save_file():1389] saving file wandb-metadata.json with policy now +2024-05-23 12:31:51,960 DEBUG HandlerThread:1833 [handler.py:handle_request():158] handle_request: python_packages +2024-05-23 12:31:51,960 DEBUG SenderThread:1833 [sender.py:send_request():405] send_request: python_packages +2024-05-23 12:31:51,962 DEBUG SenderThread:1833 [sender.py:send():378] send: telemetry +2024-05-23 12:31:51,963 DEBUG HandlerThread:1833 [handler.py:handle_request():158] handle_request: stop_status +2024-05-23 12:31:51,963 DEBUG SenderThread:1833 [sender.py:send_request():405] send_request: stop_status +2024-05-23 12:31:52,388 INFO wandb-upload_0:1833 [upload_job.py:push():130] Uploaded file /tmp/tmp0inz5gagwandb/as1pqye1-wandb-metadata.json +2024-05-23 12:31:52,544 INFO Thread-12 :1833 [dir_watcher.py:_on_file_created():271] file/dir created: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240523_123151-6w9nn173/files/wandb-metadata.json +2024-05-23 12:31:52,544 INFO Thread-12 :1833 [dir_watcher.py:_on_file_created():271] file/dir created: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240523_123151-6w9nn173/files/requirements.txt +2024-05-23 12:31:52,544 INFO Thread-12 :1833 [dir_watcher.py:_on_file_created():271] file/dir created: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240523_123151-6w9nn173/files/output.log +2024-05-23 12:31:54,543 INFO Thread-12 :1833 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240523_123151-6w9nn173/files/output.log +2024-05-23 12:31:57,078 DEBUG HandlerThread:1833 [handler.py:handle_request():158] handle_request: status_report +2024-05-23 12:32:02,551 INFO Thread-12 :1833 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240523_123151-6w9nn173/files/output.log +2024-05-23 12:32:02,582 DEBUG HandlerThread:1833 [handler.py:handle_request():158] handle_request: status_report +2024-05-23 12:32:02,892 DEBUG SenderThread:1833 [sender.py:send():378] send: exit +2024-05-23 12:32:02,892 INFO SenderThread:1833 [sender.py:send_exit():585] handling exit code: 1 +2024-05-23 12:32:02,892 INFO SenderThread:1833 [sender.py:send_exit():587] handling runtime: 11 +2024-05-23 12:32:02,894 INFO SenderThread:1833 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end +2024-05-23 12:32:02,894 INFO SenderThread:1833 [sender.py:send_exit():593] send defer +2024-05-23 12:32:02,894 DEBUG HandlerThread:1833 [handler.py:handle_request():158] handle_request: defer +2024-05-23 12:32:02,894 INFO HandlerThread:1833 [handler.py:handle_request_defer():184] handle defer: 0 +2024-05-23 12:32:02,894 DEBUG SenderThread:1833 [sender.py:send_request():405] send_request: defer +2024-05-23 12:32:02,894 INFO SenderThread:1833 [sender.py:send_request_defer():609] handle sender defer: 0 +2024-05-23 12:32:02,894 INFO SenderThread:1833 [sender.py:transition_state():613] send defer: 1 +2024-05-23 12:32:02,894 DEBUG HandlerThread:1833 [handler.py:handle_request():158] handle_request: defer +2024-05-23 12:32:02,894 INFO HandlerThread:1833 [handler.py:handle_request_defer():184] handle defer: 1 +2024-05-23 12:32:02,894 DEBUG SenderThread:1833 [sender.py:send_request():405] send_request: defer +2024-05-23 12:32:02,895 INFO SenderThread:1833 [sender.py:send_request_defer():609] handle sender defer: 1 +2024-05-23 12:32:02,895 INFO SenderThread:1833 [sender.py:transition_state():613] send defer: 2 +2024-05-23 12:32:02,895 DEBUG HandlerThread:1833 [handler.py:handle_request():158] handle_request: defer +2024-05-23 12:32:02,895 INFO HandlerThread:1833 [handler.py:handle_request_defer():184] handle defer: 2 +2024-05-23 12:32:02,895 INFO HandlerThread:1833 [system_monitor.py:finish():203] Stopping system monitor +2024-05-23 12:32:02,895 DEBUG SystemMonitor:1833 [system_monitor.py:_start():172] Starting system metrics aggregation loop +2024-05-23 12:32:02,895 DEBUG SystemMonitor:1833 [system_monitor.py:_start():179] Finished system metrics aggregation loop +2024-05-23 12:32:02,895 DEBUG SystemMonitor:1833 [system_monitor.py:_start():183] Publishing last batch of metrics +2024-05-23 12:32:02,896 INFO HandlerThread:1833 [interfaces.py:finish():200] Joined cpu monitor +2024-05-23 12:32:02,896 INFO HandlerThread:1833 [interfaces.py:finish():200] Joined disk monitor +2024-05-23 12:32:02,896 INFO HandlerThread:1833 [interfaces.py:finish():200] Joined memory monitor +2024-05-23 12:32:02,896 INFO HandlerThread:1833 [interfaces.py:finish():200] Joined network monitor +2024-05-23 12:32:02,896 DEBUG SenderThread:1833 [sender.py:send_request():405] send_request: defer +2024-05-23 12:32:02,896 INFO SenderThread:1833 [sender.py:send_request_defer():609] handle sender defer: 2 +2024-05-23 12:32:02,896 INFO SenderThread:1833 [sender.py:transition_state():613] send defer: 3 +2024-05-23 12:32:02,896 DEBUG HandlerThread:1833 [handler.py:handle_request():158] handle_request: defer +2024-05-23 12:32:02,896 INFO HandlerThread:1833 [handler.py:handle_request_defer():184] handle defer: 3 +2024-05-23 12:32:02,896 DEBUG SenderThread:1833 [sender.py:send_request():405] send_request: defer +2024-05-23 12:32:02,896 INFO SenderThread:1833 [sender.py:send_request_defer():609] handle sender defer: 3 +2024-05-23 12:32:02,897 INFO SenderThread:1833 [sender.py:transition_state():613] send defer: 4 +2024-05-23 12:32:02,897 DEBUG HandlerThread:1833 [handler.py:handle_request():158] handle_request: defer +2024-05-23 12:32:02,897 INFO HandlerThread:1833 [handler.py:handle_request_defer():184] handle defer: 4 +2024-05-23 12:32:02,897 DEBUG SenderThread:1833 [sender.py:send_request():405] send_request: defer +2024-05-23 12:32:02,897 INFO SenderThread:1833 [sender.py:send_request_defer():609] handle sender defer: 4 +2024-05-23 12:32:02,897 INFO SenderThread:1833 [sender.py:transition_state():613] send defer: 5 +2024-05-23 12:32:02,897 DEBUG HandlerThread:1833 [handler.py:handle_request():158] handle_request: defer +2024-05-23 12:32:02,897 INFO HandlerThread:1833 [handler.py:handle_request_defer():184] handle defer: 5 +2024-05-23 12:32:02,897 DEBUG SenderThread:1833 [sender.py:send():378] send: summary +2024-05-23 12:32:02,898 INFO SenderThread:1833 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end +2024-05-23 12:32:02,898 DEBUG SenderThread:1833 [sender.py:send_request():405] send_request: defer +2024-05-23 12:32:02,898 INFO SenderThread:1833 [sender.py:send_request_defer():609] handle sender defer: 5 +2024-05-23 12:32:02,898 INFO SenderThread:1833 [sender.py:transition_state():613] send defer: 6 +2024-05-23 12:32:02,898 DEBUG HandlerThread:1833 [handler.py:handle_request():158] handle_request: defer +2024-05-23 12:32:02,898 INFO HandlerThread:1833 [handler.py:handle_request_defer():184] handle defer: 6 +2024-05-23 12:32:02,898 DEBUG SenderThread:1833 [sender.py:send_request():405] send_request: defer +2024-05-23 12:32:02,898 INFO SenderThread:1833 [sender.py:send_request_defer():609] handle sender defer: 6 +2024-05-23 12:32:02,903 DEBUG HandlerThread:1833 [handler.py:handle_request():158] handle_request: status_report +2024-05-23 12:32:02,970 INFO SenderThread:1833 [sender.py:transition_state():613] send defer: 7 +2024-05-23 12:32:02,970 DEBUG HandlerThread:1833 [handler.py:handle_request():158] handle_request: defer +2024-05-23 12:32:02,970 INFO HandlerThread:1833 [handler.py:handle_request_defer():184] handle defer: 7 +2024-05-23 12:32:02,970 DEBUG SenderThread:1833 [sender.py:send_request():405] send_request: defer +2024-05-23 12:32:02,970 INFO SenderThread:1833 [sender.py:send_request_defer():609] handle sender defer: 7 +2024-05-23 12:32:03,554 INFO Thread-12 :1833 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240523_123151-6w9nn173/files/config.yaml +2024-05-23 12:32:03,554 INFO Thread-12 :1833 [dir_watcher.py:_on_file_created():271] file/dir created: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240523_123151-6w9nn173/files/wandb-summary.json +2024-05-23 12:32:03,892 DEBUG HandlerThread:1833 [handler.py:handle_request():158] handle_request: poll_exit +2024-05-23 12:32:04,098 INFO SenderThread:1833 [sender.py:transition_state():613] send defer: 8 +2024-05-23 12:32:04,098 DEBUG SenderThread:1833 [sender.py:send_request():405] send_request: poll_exit +2024-05-23 12:32:04,098 DEBUG HandlerThread:1833 [handler.py:handle_request():158] handle_request: defer +2024-05-23 12:32:04,098 INFO HandlerThread:1833 [handler.py:handle_request_defer():184] handle defer: 8 +2024-05-23 12:32:04,098 DEBUG SenderThread:1833 [sender.py:send_request():405] send_request: defer +2024-05-23 12:32:04,098 INFO SenderThread:1833 [sender.py:send_request_defer():609] handle sender defer: 8 +2024-05-23 12:32:04,098 INFO SenderThread:1833 [job_builder.py:build():432] Attempting to build job artifact +2024-05-23 12:32:04,099 INFO SenderThread:1833 [job_builder.py:_get_source_type():576] no source found +2024-05-23 12:32:04,099 INFO SenderThread:1833 [sender.py:transition_state():613] send defer: 9 +2024-05-23 12:32:04,099 DEBUG HandlerThread:1833 [handler.py:handle_request():158] handle_request: defer +2024-05-23 12:32:04,099 INFO HandlerThread:1833 [handler.py:handle_request_defer():184] handle defer: 9 +2024-05-23 12:32:04,099 DEBUG SenderThread:1833 [sender.py:send_request():405] send_request: defer +2024-05-23 12:32:04,099 INFO SenderThread:1833 [sender.py:send_request_defer():609] handle sender defer: 9 +2024-05-23 12:32:04,099 INFO SenderThread:1833 [dir_watcher.py:finish():358] shutting down directory watcher +2024-05-23 12:32:04,556 INFO SenderThread:1833 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240523_123151-6w9nn173/files/output.log +2024-05-23 12:32:04,556 INFO SenderThread:1833 [dir_watcher.py:finish():388] scan: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240523_123151-6w9nn173/files +2024-05-23 12:32:04,556 INFO SenderThread:1833 [dir_watcher.py:finish():402] scan save: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240523_123151-6w9nn173/files/requirements.txt requirements.txt +2024-05-23 12:32:04,556 INFO SenderThread:1833 [dir_watcher.py:finish():402] scan save: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240523_123151-6w9nn173/files/config.yaml config.yaml +2024-05-23 12:32:04,559 INFO SenderThread:1833 [dir_watcher.py:finish():402] scan save: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240523_123151-6w9nn173/files/wandb-metadata.json wandb-metadata.json +2024-05-23 12:32:04,559 INFO SenderThread:1833 [dir_watcher.py:finish():402] scan save: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240523_123151-6w9nn173/files/output.log output.log +2024-05-23 12:32:04,561 INFO SenderThread:1833 [dir_watcher.py:finish():402] scan save: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240523_123151-6w9nn173/files/wandb-summary.json wandb-summary.json +2024-05-23 12:32:04,561 INFO SenderThread:1833 [sender.py:transition_state():613] send defer: 10 +2024-05-23 12:32:04,562 DEBUG HandlerThread:1833 [handler.py:handle_request():158] handle_request: defer +2024-05-23 12:32:04,562 INFO HandlerThread:1833 [handler.py:handle_request_defer():184] handle defer: 10 +2024-05-23 12:32:04,562 DEBUG SenderThread:1833 [sender.py:send_request():405] send_request: defer +2024-05-23 12:32:04,562 INFO SenderThread:1833 [sender.py:send_request_defer():609] handle sender defer: 10 +2024-05-23 12:32:04,562 INFO SenderThread:1833 [file_pusher.py:finish():169] shutting down file pusher +2024-05-23 12:32:04,797 INFO wandb-upload_0:1833 [upload_job.py:push():130] Uploaded file /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240523_123151-6w9nn173/files/requirements.txt +2024-05-23 12:32:04,892 DEBUG HandlerThread:1833 [handler.py:handle_request():158] handle_request: poll_exit +2024-05-23 12:32:04,892 DEBUG SenderThread:1833 [sender.py:send_request():405] send_request: poll_exit +2024-05-23 12:32:05,174 INFO wandb-upload_3:1833 [upload_job.py:push():130] Uploaded file /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240523_123151-6w9nn173/files/wandb-summary.json +2024-05-23 12:32:05,192 INFO wandb-upload_2:1833 [upload_job.py:push():130] Uploaded file /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240523_123151-6w9nn173/files/output.log +2024-05-23 12:32:05,204 INFO wandb-upload_1:1833 [upload_job.py:push():130] Uploaded file /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240523_123151-6w9nn173/files/config.yaml +2024-05-23 12:32:05,404 INFO Thread-11 (_thread_body):1833 [sender.py:transition_state():613] send defer: 11 +2024-05-23 12:32:05,404 DEBUG HandlerThread:1833 [handler.py:handle_request():158] handle_request: defer +2024-05-23 12:32:05,404 INFO HandlerThread:1833 [handler.py:handle_request_defer():184] handle defer: 11 +2024-05-23 12:32:05,404 DEBUG SenderThread:1833 [sender.py:send_request():405] send_request: defer +2024-05-23 12:32:05,404 INFO SenderThread:1833 [sender.py:send_request_defer():609] handle sender defer: 11 +2024-05-23 12:32:05,404 INFO SenderThread:1833 [file_pusher.py:join():175] waiting for file pusher +2024-05-23 12:32:05,404 INFO SenderThread:1833 [sender.py:transition_state():613] send defer: 12 +2024-05-23 12:32:05,405 DEBUG HandlerThread:1833 [handler.py:handle_request():158] handle_request: defer +2024-05-23 12:32:05,405 INFO HandlerThread:1833 [handler.py:handle_request_defer():184] handle defer: 12 +2024-05-23 12:32:05,405 DEBUG SenderThread:1833 [sender.py:send_request():405] send_request: defer +2024-05-23 12:32:05,405 INFO SenderThread:1833 [sender.py:send_request_defer():609] handle sender defer: 12 +2024-05-23 12:32:05,405 INFO SenderThread:1833 [file_stream.py:finish():601] file stream finish called +2024-05-23 12:32:05,588 INFO SenderThread:1833 [file_stream.py:finish():605] file stream finish is done +2024-05-23 12:32:05,588 INFO SenderThread:1833 [sender.py:transition_state():613] send defer: 13 +2024-05-23 12:32:05,588 DEBUG HandlerThread:1833 [handler.py:handle_request():158] handle_request: defer +2024-05-23 12:32:05,588 INFO HandlerThread:1833 [handler.py:handle_request_defer():184] handle defer: 13 +2024-05-23 12:32:05,588 DEBUG SenderThread:1833 [sender.py:send_request():405] send_request: defer +2024-05-23 12:32:05,588 INFO SenderThread:1833 [sender.py:send_request_defer():609] handle sender defer: 13 +2024-05-23 12:32:05,588 INFO SenderThread:1833 [sender.py:transition_state():613] send defer: 14 +2024-05-23 12:32:05,588 DEBUG HandlerThread:1833 [handler.py:handle_request():158] handle_request: defer +2024-05-23 12:32:05,588 INFO HandlerThread:1833 [handler.py:handle_request_defer():184] handle defer: 14 +2024-05-23 12:32:05,589 DEBUG SenderThread:1833 [sender.py:send():378] send: final +2024-05-23 12:32:05,589 DEBUG SenderThread:1833 [sender.py:send():378] send: footer +2024-05-23 12:32:05,589 DEBUG SenderThread:1833 [sender.py:send_request():405] send_request: defer +2024-05-23 12:32:05,589 INFO SenderThread:1833 [sender.py:send_request_defer():609] handle sender defer: 14 +2024-05-23 12:32:05,589 DEBUG HandlerThread:1833 [handler.py:handle_request():158] handle_request: poll_exit +2024-05-23 12:32:05,590 DEBUG HandlerThread:1833 [handler.py:handle_request():158] handle_request: poll_exit +2024-05-23 12:32:05,590 DEBUG HandlerThread:1833 [handler.py:handle_request():158] handle_request: server_info +2024-05-23 12:32:05,590 DEBUG HandlerThread:1833 [handler.py:handle_request():158] handle_request: get_summary +2024-05-23 12:32:05,590 DEBUG HandlerThread:1833 [handler.py:handle_request():158] handle_request: sampled_history +2024-05-23 12:32:05,590 DEBUG HandlerThread:1833 [handler.py:handle_request():158] handle_request: internal_messages +2024-05-23 12:32:05,590 DEBUG SenderThread:1833 [sender.py:send_request():405] send_request: poll_exit +2024-05-23 12:32:05,590 DEBUG SenderThread:1833 [sender.py:send_request():405] send_request: poll_exit +2024-05-23 12:32:05,590 DEBUG SenderThread:1833 [sender.py:send_request():405] send_request: server_info +2024-05-23 12:32:05,654 INFO MainThread:1833 [wandb_run.py:_footer_history_summary_info():3994] rendering history +2024-05-23 12:32:05,654 INFO MainThread:1833 [wandb_run.py:_footer_history_summary_info():4026] rendering summary +2024-05-23 12:32:05,654 INFO MainThread:1833 [wandb_run.py:_footer_sync_info():3953] logging synced files +2024-05-23 12:32:05,654 DEBUG HandlerThread:1833 [handler.py:handle_request():158] handle_request: shutdown +2024-05-23 12:32:05,654 INFO HandlerThread:1833 [handler.py:finish():882] shutting down handler +2024-05-23 12:32:06,590 INFO WriterThread:1833 [datastore.py:close():296] close: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240523_123151-6w9nn173/run-6w9nn173.wandb +2024-05-23 12:32:06,654 INFO SenderThread:1833 [sender.py:finish():1545] shutting down sender +2024-05-23 12:32:06,654 INFO SenderThread:1833 [file_pusher.py:finish():169] shutting down file pusher +2024-05-23 12:32:06,654 INFO SenderThread:1833 [file_pusher.py:join():175] waiting for file pusher diff --git a/lm-evaluation-harness/wandb/run-20240523_123151-6w9nn173/logs/debug.log b/lm-evaluation-harness/wandb/run-20240523_123151-6w9nn173/logs/debug.log new file mode 100644 index 0000000000000000000000000000000000000000..61cd749dfca192731861c93cd69d11458822cd44 --- /dev/null +++ b/lm-evaluation-harness/wandb/run-20240523_123151-6w9nn173/logs/debug.log @@ -0,0 +1,29 @@ +2024-05-23 12:31:51,173 INFO MainThread:1678 [wandb_setup.py:_flush():76] Current SDK version is 0.17.0 +2024-05-23 12:31:51,173 INFO MainThread:1678 [wandb_setup.py:_flush():76] Configure stats pid to 1678 +2024-05-23 12:31:51,173 INFO MainThread:1678 [wandb_setup.py:_flush():76] Loading settings from /root/.config/wandb/settings +2024-05-23 12:31:51,173 INFO MainThread:1678 [wandb_setup.py:_flush():76] Loading settings from /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/settings +2024-05-23 12:31:51,173 INFO MainThread:1678 [wandb_setup.py:_flush():76] Loading settings from environment variables: {} +2024-05-23 12:31:51,173 INFO MainThread:1678 [wandb_setup.py:_flush():76] Applying setup settings: {'_disable_service': False} +2024-05-23 12:31:51,173 WARNING MainThread:1678 [wandb_setup.py:_flush():76] Could not find program at -m lm_eval.__main__ +2024-05-23 12:31:51,173 INFO MainThread:1678 [wandb_setup.py:_flush():76] Inferring run settings from compute environment: {'program_relpath': None, 'program': '-m lm_eval.__main__'} +2024-05-23 12:31:51,173 INFO MainThread:1678 [wandb_setup.py:_flush():76] Applying login settings: {} +2024-05-23 12:31:51,173 INFO MainThread:1678 [wandb_init.py:_log_setup():520] Logging user logs to /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240523_123151-6w9nn173/logs/debug.log +2024-05-23 12:31:51,173 INFO MainThread:1678 [wandb_init.py:_log_setup():521] Logging internal logs to /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240523_123151-6w9nn173/logs/debug-internal.log +2024-05-23 12:31:51,174 INFO MainThread:1678 [wandb_init.py:init():560] calling init triggers +2024-05-23 12:31:51,174 INFO MainThread:1678 [wandb_init.py:init():567] wandb.init called with sweep_config: {} +config: {} +2024-05-23 12:31:51,174 INFO MainThread:1678 [wandb_init.py:init():610] starting backend +2024-05-23 12:31:51,174 INFO MainThread:1678 [wandb_init.py:init():614] setting up manager +2024-05-23 12:31:51,177 INFO MainThread:1678 [backend.py:_multiprocessing_setup():105] multiprocessing start_methods=fork,spawn,forkserver, using: spawn +2024-05-23 12:31:51,178 INFO MainThread:1678 [wandb_init.py:init():622] backend started and connected +2024-05-23 12:31:51,182 INFO MainThread:1678 [wandb_init.py:init():711] updated telemetry +2024-05-23 12:31:51,189 INFO MainThread:1678 [wandb_init.py:init():744] communicating run to backend with 90.0 second timeout +2024-05-23 12:31:51,545 INFO MainThread:1678 [wandb_run.py:_on_init():2396] communicating current version +2024-05-23 12:31:51,656 INFO MainThread:1678 [wandb_run.py:_on_init():2405] got version response +2024-05-23 12:31:51,656 INFO MainThread:1678 [wandb_init.py:init():795] starting run threads in backend +2024-05-23 12:31:51,961 INFO MainThread:1678 [wandb_run.py:_console_start():2374] atexit reg +2024-05-23 12:31:51,961 INFO MainThread:1678 [wandb_run.py:_redirect():2229] redirect: wrap_raw +2024-05-23 12:31:51,961 INFO MainThread:1678 [wandb_run.py:_redirect():2294] Wrapping output streams. +2024-05-23 12:31:51,961 INFO MainThread:1678 [wandb_run.py:_redirect():2319] Redirects installed. +2024-05-23 12:31:51,963 INFO MainThread:1678 [wandb_init.py:init():838] run started, returning control to user process +2024-05-23 12:32:06,655 WARNING MsgRouterThr:1678 [router.py:message_loop():77] message_loop has been closed diff --git a/lm-evaluation-harness/wandb/run-20240523_123151-6w9nn173/run-6w9nn173.wandb b/lm-evaluation-harness/wandb/run-20240523_123151-6w9nn173/run-6w9nn173.wandb new file mode 100644 index 0000000000000000000000000000000000000000..71ac22284abb51467b395a81edaff20e9a28e8ee Binary files /dev/null and b/lm-evaluation-harness/wandb/run-20240523_123151-6w9nn173/run-6w9nn173.wandb differ diff --git a/lm-evaluation-harness/wandb/run-20240523_130610-knsm0gsf/files/config.yaml b/lm-evaluation-harness/wandb/run-20240523_130610-knsm0gsf/files/config.yaml new file mode 100644 index 0000000000000000000000000000000000000000..bb7849bca1c68c27e07096cbd12097cd995898ca --- /dev/null +++ b/lm-evaluation-harness/wandb/run-20240523_130610-knsm0gsf/files/config.yaml @@ -0,0 +1,43 @@ +wandb_version: 1 + +_wandb: + desc: null + value: + python_version: 3.10.12 + cli_version: 0.17.0 + framework: huggingface + huggingface_version: 4.41.1 + is_jupyter_run: false + is_kaggle_kernel: false + start_time: 1716469570 + t: + 1: + - 1 + - 5 + - 11 + - 49 + - 51 + - 53 + - 55 + - 71 + - 98 + - 100 + 2: + - 1 + - 5 + - 11 + - 49 + - 51 + - 53 + - 55 + - 71 + - 98 + - 100 + 3: + - 23 + 4: 3.10.12 + 5: 0.17.0 + 6: 4.41.1 + 8: + - 5 + 13: linux-x86_64 diff --git a/lm-evaluation-harness/wandb/run-20240523_130610-knsm0gsf/files/output.log b/lm-evaluation-harness/wandb/run-20240523_130610-knsm0gsf/files/output.log new file mode 100644 index 0000000000000000000000000000000000000000..424e46950ad7c294d5ae63b2f84198618eb89c00 --- /dev/null +++ b/lm-evaluation-harness/wandb/run-20240523_130610-knsm0gsf/files/output.log @@ -0,0 +1,34 @@ + +2024-05-23:13:06:10,742 INFO [__main__.py:251] Verbosity set to INFO +2024-05-23:13:06:19,245 INFO [__main__.py:335] Selected Tasks: ['arc_easy', 'hellaswag', 'mrpc', 'openbookqa', 'sst2', 'winogrande'] +2024-05-23:13:06:19,246 INFO [evaluator.py:131] Setting random seed to 0 | Setting numpy seed to 1234 | Setting torch manual seed to 1234 +2024-05-23:13:06:19,246 INFO [evaluator.py:177] Initializing hf model, with arguments: {'pretrained': '/mnt/weka/peacock/experiments/llama/checkpoint/llamav2-3b//hf_ckpt//global_step18000'} +2024-05-23:13:06:21,516 INFO [huggingface.py:164] Using device 'cuda' +Traceback (most recent call last): + File "/usr/lib/python3.10/runpy.py", line 196, in _run_module_as_main + return _run_code(code, main_globals, None, + File "/usr/lib/python3.10/runpy.py", line 86, in _run_code + exec(code, run_globals) + File "/mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/lm_eval/__main__.py", line 417, in + cli_evaluate() + File "/mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/lm_eval/__main__.py", line 341, in cli_evaluate + results = evaluator.simple_evaluate( + File "/mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/lm_eval/utils.py", line 288, in _wrapper + return fn(*args, **kwargs) + File "/mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/lm_eval/evaluator.py", line 180, in simple_evaluate + lm = lm_eval.api.registry.get_model(model).create_from_arg_string( + File "/mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/lm_eval/api/model.py", line 134, in create_from_arg_string + return cls(**args, **args2) + File "/mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/lm_eval/models/huggingface.py", line 190, in __init__ + self._get_config( + File "/mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/lm_eval/models/huggingface.py", line 471, in _get_config + self._config = transformers.AutoConfig.from_pretrained( + File "/usr/local/lib/python3.10/dist-packages/transformers/models/auto/configuration_auto.py", line 934, in from_pretrained + config_dict, unused_kwargs = PretrainedConfig.get_config_dict(pretrained_model_name_or_path, **kwargs) + File "/usr/local/lib/python3.10/dist-packages/transformers/configuration_utils.py", line 632, in get_config_dict + config_dict, kwargs = cls._get_config_dict(pretrained_model_name_or_path, **kwargs) + File "/usr/local/lib/python3.10/dist-packages/transformers/configuration_utils.py", line 689, in _get_config_dict + resolved_config_file = cached_file( + File "/usr/local/lib/python3.10/dist-packages/transformers/utils/hub.py", line 370, in cached_file + raise EnvironmentError( +OSError: /mnt/weka/peacock/experiments/llama/checkpoint/llamav2-3b//hf_ckpt//global_step18000 does not appear to have a file named config.json. Checkout 'https://huggingface.co//mnt/weka/peacock/experiments/llama/checkpoint/llamav2-3b//hf_ckpt//global_step18000/tree/main' for available files. \ No newline at end of file diff --git a/lm-evaluation-harness/wandb/run-20240523_130610-knsm0gsf/files/requirements.txt b/lm-evaluation-harness/wandb/run-20240523_130610-knsm0gsf/files/requirements.txt new file mode 100644 index 0000000000000000000000000000000000000000..f675c3016b5332c1acf28f436e0b60adeead9c12 --- /dev/null +++ b/lm-evaluation-harness/wandb/run-20240523_130610-knsm0gsf/files/requirements.txt @@ -0,0 +1,155 @@ +DataProperty==1.0.1 +GitPython==3.1.43 +Jinja2==3.1.4 +Markdown==3.6 +MarkupSafe==2.1.5 +Pillow-SIMD==7.0.0.post3 +PyYAML==6.0 +Werkzeug==3.0.3 +absl-py==2.1.0 +accelerate==0.30.1 +aiohttp==3.9.5 +aiosignal==1.3.1 +async-timeout==4.0.3 +attrs==23.2.0 +av==9.2.0 +cachetools==5.3.3 +certifi==2024.2.2 +cffi==1.15.1 +cfgv==3.4.0 +chardet==5.2.0 +charset-normalizer==3.3.2 +click==8.1.7 +cmake==3.29.2 +colorama==0.4.6 +datasets==2.19.1 +deepspeed==0.12.4+hpu.synapse.v1.15.1 +dill==0.3.8 +distlib==0.3.8 +docker-pycreds==0.4.0 +einops==0.8.0 +evaluate==0.4.2 +exceptiongroup==1.2.1 +expecttest==0.2.1 +filelock==3.14.0 +frozenlist==1.4.1 +fsspec==2024.3.1 +gitdb==4.0.11 +google-auth-oauthlib==0.4.6 +google-auth==2.29.0 +grpcio==1.63.0 +habana-media-loader==1.15.1.15 +habana-pyhlml==1.15.1.15 +habana-torch-dataloader==1.15.1.15 +habana-torch-plugin==1.15.1.15 +habana_gpu_migration==1.15.1.15 +habana_quantization_toolkit==1.15.1.15 +hjson==3.1.0 +huggingface-hub==0.23.1 +identify==2.5.36 +idna==3.7 +iniconfig==2.0.0 +joblib==1.4.2 +jsonlines==4.0.0 +lightning-habana==1.4.0 +lightning-utilities==0.11.2 +lightning==2.2.0.post0 +lm_eval==0.4.2 +lm_eval==0.4.2 +lm_eval==0.4.2 +lxml==5.2.2 +mbstrdecoder==1.1.3 +more-itertools==10.2.0 +mpi4py==3.1.4 +mpmath==1.3.0 +multidict==6.0.5 +multiprocess==0.70.16 +networkx==3.3 +ninja==1.11.1.1 +nltk==3.8.1 +nodeenv==1.8.0 +numexpr==2.10.0 +numpy==1.23.5 +oauthlib==3.2.2 +packaging==24.0 +pandas==2.0.1 +pathspec==0.12.1 +pathvalidate==3.2.0 +peft==0.11.1 +perfetto==0.7.0 +pillow==10.3.0 +pip==22.0.2 +pip==23.3.1 +platformdirs==4.2.1 +pluggy==1.5.0 +portalocker==2.8.2 +pre-commit==3.3.3 +pretty-errors==1.2.25 +protobuf==3.20.3 +psutil==5.9.8 +py-cpuinfo==9.0.0 +pyarrow-hotfix==0.6 +pyarrow==16.1.0 +pyasn1==0.6.0 +pyasn1_modules==0.4.0 +pybind11==2.10.4 +pycparser==2.22 +pydantic==1.10.13 +pynvml==8.0.4 +pytablewriter==1.2.0 +pytest==8.2.0 +python-dateutil==2.9.0.post0 +pytorch-lightning==2.2.4 +pytz==2024.1 +regex==2023.5.5 +requests-oauthlib==2.0.0 +requests==2.31.0 +rouge_score==0.1.2 +rsa==4.9 +sacrebleu==2.4.2 +safetensors==0.4.3 +scikit-learn==1.5.0 +scipy==1.13.1 +sentencepiece==0.2.0 +sentry-sdk==2.3.0 +setproctitle==1.3.3 +setuptools==59.6.0 +setuptools==69.5.1 +six==1.16.0 +smmap==5.0.1 +sqlitedict==2.1.0 +symengine==0.11.0 +sympy==1.12 +tabledata==1.3.3 +tabulate==0.9.0 +tcolorpy==0.1.6 +tdqm==0.0.1 +tensorboard-data-server==0.6.1 +tensorboard-plugin-wit==1.8.1 +tensorboard==2.11.2 +threadpoolctl==3.5.0 +tokenizers==0.19.1 +tomli==2.0.1 +torch==2.2.0a0+git8964477 +torch_tb_profiler==0.4.0 +torchaudio==2.2.0+08901ad +torchdata==0.7.1+5e6f7b7 +torchmetrics==1.4.0 +torchtext==0.17.0+400da5c +torchvision==0.17.0+b2383d4 +tqdm-multiprocess==0.0.11 +tqdm==4.66.4 +transformers==4.41.1 +typepy==1.3.2 +typing_extensions==4.11.0 +tzdata==2024.1 +urllib3==1.26.18 +virtualenv==20.26.1 +wandb==0.17.0 +wheel==0.37.1 +wheel==0.43.0 +word2number==1.1 +xxhash==3.4.1 +yamllint==1.35.1 +yarl==1.9.4 +zstandard==0.22.0 \ No newline at end of file diff --git a/lm-evaluation-harness/wandb/run-20240523_130610-knsm0gsf/files/wandb-metadata.json b/lm-evaluation-harness/wandb/run-20240523_130610-knsm0gsf/files/wandb-metadata.json new file mode 100644 index 0000000000000000000000000000000000000000..bfc5c56bf1b67b26789956407c8906810f4b47c5 --- /dev/null +++ b/lm-evaluation-harness/wandb/run-20240523_130610-knsm0gsf/files/wandb-metadata.json @@ -0,0 +1,850 @@ +{ + "os": "Linux-5.15.0-92-generic-x86_64-with-glibc2.35", + "python": "3.10.12", + "heartbeatAt": "2024-05-23T13:06:10.509166", + "startedAt": "2024-05-23T13:06:10.025312", + "docker": null, + "cuda": null, + "args": [ + "--model", + "hf", + "--model_args", + "pretrained=/mnt/weka/peacock/experiments/llama/checkpoint/llamav2-3b//hf_ckpt//global_step18000", + "--tasks", + "hellaswag,arc_easy,openbookqa,winogrande,sst2,mrpc", + "--batch_size", + "auto", + "--wandb_args", + "project=bharatgpt,group=trial_expt_2" + ], + "state": "running", + "program": "-m lm_eval.__main__", + "codePathLocal": null, + "git": { + "remote": "https://github.com/EleutherAI/lm-evaluation-harness", + "commit": null + }, + "email": null, + "root": "/mnt/weka/peacock/idc/cronscript/lm-evaluation-harness", + "host": "peacock-evaluation-worker-0", + "username": "root", + "executable": "/usr/bin/python3", + "cpu_count": 80, + "cpu_count_logical": 160, + "cpu_freq": { + "current": 2327.4999875000003, + "min": 800.0, + "max": 3400.0 + }, + "cpu_freq_per_core": [ + { + "current": 3390.951, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3390.951, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.002, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3390.945, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + } + ], + "disk": { + "/": { + "total": 877.6341285705566, + "used": 211.6283836364746 + } + }, + "memory": { + "total": 1007.4379539489746 + } +} diff --git a/lm-evaluation-harness/wandb/run-20240523_130610-knsm0gsf/files/wandb-summary.json b/lm-evaluation-harness/wandb/run-20240523_130610-knsm0gsf/files/wandb-summary.json new file mode 100644 index 0000000000000000000000000000000000000000..8bf99d152ad35c3699ec8600ecb8b169d4e35875 --- /dev/null +++ b/lm-evaluation-harness/wandb/run-20240523_130610-knsm0gsf/files/wandb-summary.json @@ -0,0 +1 @@ +{"_wandb": {"runtime": 11}} \ No newline at end of file diff --git a/lm-evaluation-harness/wandb/run-20240523_130610-knsm0gsf/logs/debug-internal.log b/lm-evaluation-harness/wandb/run-20240523_130610-knsm0gsf/logs/debug-internal.log new file mode 100644 index 0000000000000000000000000000000000000000..7d9c67db4c0154d0ea0d56f3e2e427ff14a76374 --- /dev/null +++ b/lm-evaluation-harness/wandb/run-20240523_130610-knsm0gsf/logs/debug-internal.log @@ -0,0 +1,183 @@ +2024-05-23 13:06:10,048 INFO StreamThr :2100 [internal.py:wandb_internal():85] W&B internal server running at pid: 2100, started at: 2024-05-23 13:06:10.045985 +2024-05-23 13:06:10,052 DEBUG HandlerThread:2100 [handler.py:handle_request():158] handle_request: status +2024-05-23 13:06:10,053 INFO WriterThread:2100 [datastore.py:open_for_write():87] open: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240523_130610-knsm0gsf/run-knsm0gsf.wandb +2024-05-23 13:06:10,056 DEBUG SenderThread:2100 [sender.py:send():378] send: header +2024-05-23 13:06:10,059 DEBUG SenderThread:2100 [sender.py:send():378] send: run +2024-05-23 13:06:10,313 INFO SenderThread:2100 [dir_watcher.py:__init__():211] watching files in: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240523_130610-knsm0gsf/files +2024-05-23 13:06:10,313 INFO SenderThread:2100 [sender.py:_start_run_threads():1123] run started: knsm0gsf with start time 1716469570.04584 +2024-05-23 13:06:10,316 DEBUG HandlerThread:2100 [handler.py:handle_request():158] handle_request: check_version +2024-05-23 13:06:10,317 DEBUG SenderThread:2100 [sender.py:send_request():405] send_request: check_version +2024-05-23 13:06:10,432 DEBUG HandlerThread:2100 [handler.py:handle_request():158] handle_request: run_start +2024-05-23 13:06:10,434 DEBUG HandlerThread:2100 [system_info.py:__init__():26] System info init +2024-05-23 13:06:10,434 DEBUG HandlerThread:2100 [system_info.py:__init__():41] System info init done +2024-05-23 13:06:10,434 INFO HandlerThread:2100 [system_monitor.py:start():194] Starting system monitor +2024-05-23 13:06:10,435 INFO SystemMonitor:2100 [system_monitor.py:_start():158] Starting system asset monitoring threads +2024-05-23 13:06:10,435 INFO HandlerThread:2100 [system_monitor.py:probe():214] Collecting system info +2024-05-23 13:06:10,441 INFO SystemMonitor:2100 [interfaces.py:start():188] Started cpu monitoring +2024-05-23 13:06:10,447 INFO SystemMonitor:2100 [interfaces.py:start():188] Started disk monitoring +2024-05-23 13:06:10,447 INFO SystemMonitor:2100 [interfaces.py:start():188] Started memory monitoring +2024-05-23 13:06:10,447 INFO SystemMonitor:2100 [interfaces.py:start():188] Started network monitoring +2024-05-23 13:06:10,509 DEBUG HandlerThread:2100 [system_info.py:probe():150] Probing system +2024-05-23 13:06:10,512 DEBUG HandlerThread:2100 [system_info.py:_probe_git():135] Probing git +2024-05-23 13:06:10,522 ERROR HandlerThread:2100 [gitlib.py:root():92] git root error: Cmd('git') failed due to: exit code(128) + cmdline: git rev-parse --show-toplevel + stderr: 'fatal: detected dubious ownership in repository at '/mnt/weka/peacock/idc/cronscript/lm-evaluation-harness' +To add an exception for this directory, call: + + git config --global --add safe.directory /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness' +2024-05-23 13:06:10,522 DEBUG HandlerThread:2100 [system_info.py:_probe_git():143] Probing git done +2024-05-23 13:06:10,522 DEBUG HandlerThread:2100 [system_info.py:probe():198] Probing system done +2024-05-23 13:06:10,522 DEBUG HandlerThread:2100 [system_monitor.py:probe():223] {'os': 'Linux-5.15.0-92-generic-x86_64-with-glibc2.35', 'python': '3.10.12', 'heartbeatAt': '2024-05-23T13:06:10.509166', 'startedAt': '2024-05-23T13:06:10.025312', 'docker': None, 'cuda': None, 'args': ('--model', 'hf', '--model_args', 'pretrained=/mnt/weka/peacock/experiments/llama/checkpoint/llamav2-3b//hf_ckpt//global_step18000', '--tasks', 'hellaswag,arc_easy,openbookqa,winogrande,sst2,mrpc', '--batch_size', 'auto', '--wandb_args', 'project=bharatgpt,group=trial_expt_2'), 'state': 'running', 'program': '-m lm_eval.__main__', 'codePathLocal': None, 'git': {'remote': 'https://github.com/EleutherAI/lm-evaluation-harness', 'commit': None}, 'email': None, 'root': '/mnt/weka/peacock/idc/cronscript/lm-evaluation-harness', 'host': 'peacock-evaluation-worker-0', 'username': 'root', 'executable': '/usr/bin/python3', 'cpu_count': 80, 'cpu_count_logical': 160, 'cpu_freq': {'current': 2327.4999875000003, 'min': 800.0, 'max': 3400.0}, 'cpu_freq_per_core': [{'current': 3390.951, 'min': 800.0, 'max': 3400.0}, {'current': 3390.951, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.002, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 3390.945, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}], 'disk': {'/': {'total': 877.6341285705566, 'used': 211.6283836364746}}, 'memory': {'total': 1007.4379539489746}} +2024-05-23 13:06:10,522 INFO HandlerThread:2100 [system_monitor.py:probe():224] Finished collecting system info +2024-05-23 13:06:10,522 INFO HandlerThread:2100 [system_monitor.py:probe():227] Publishing system info +2024-05-23 13:06:10,525 INFO HandlerThread:2100 [system_monitor.py:probe():229] Finished publishing system info +2024-05-23 13:06:10,530 DEBUG SenderThread:2100 [sender.py:send():378] send: files +2024-05-23 13:06:10,530 INFO SenderThread:2100 [sender.py:_save_file():1389] saving file wandb-metadata.json with policy now +2024-05-23 13:06:10,710 DEBUG HandlerThread:2100 [handler.py:handle_request():158] handle_request: python_packages +2024-05-23 13:06:10,711 DEBUG SenderThread:2100 [sender.py:send_request():405] send_request: python_packages +2024-05-23 13:06:10,711 DEBUG HandlerThread:2100 [handler.py:handle_request():158] handle_request: stop_status +2024-05-23 13:06:10,721 DEBUG SenderThread:2100 [sender.py:send_request():405] send_request: stop_status +2024-05-23 13:06:10,869 DEBUG SenderThread:2100 [sender.py:send():378] send: telemetry +2024-05-23 13:06:11,134 INFO wandb-upload_0:2100 [upload_job.py:push():130] Uploaded file /tmp/tmpnckw1havwandb/vo13uadd-wandb-metadata.json +2024-05-23 13:06:11,316 INFO Thread-12 :2100 [dir_watcher.py:_on_file_created():271] file/dir created: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240523_130610-knsm0gsf/files/output.log +2024-05-23 13:06:11,316 INFO Thread-12 :2100 [dir_watcher.py:_on_file_created():271] file/dir created: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240523_130610-knsm0gsf/files/requirements.txt +2024-05-23 13:06:11,316 INFO Thread-12 :2100 [dir_watcher.py:_on_file_created():271] file/dir created: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240523_130610-knsm0gsf/files/wandb-metadata.json +2024-05-23 13:06:13,315 INFO Thread-12 :2100 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240523_130610-knsm0gsf/files/output.log +2024-05-23 13:06:15,874 DEBUG HandlerThread:2100 [handler.py:handle_request():158] handle_request: status_report +2024-05-23 13:06:21,247 DEBUG HandlerThread:2100 [handler.py:handle_request():158] handle_request: status_report +2024-05-23 13:06:21,327 INFO Thread-12 :2100 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240523_130610-knsm0gsf/files/output.log +2024-05-23 13:06:21,526 DEBUG SenderThread:2100 [sender.py:send():378] send: exit +2024-05-23 13:06:21,526 INFO SenderThread:2100 [sender.py:send_exit():585] handling exit code: 1 +2024-05-23 13:06:21,527 INFO SenderThread:2100 [sender.py:send_exit():587] handling runtime: 11 +2024-05-23 13:06:21,528 INFO SenderThread:2100 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end +2024-05-23 13:06:21,528 INFO SenderThread:2100 [sender.py:send_exit():593] send defer +2024-05-23 13:06:21,529 DEBUG HandlerThread:2100 [handler.py:handle_request():158] handle_request: defer +2024-05-23 13:06:21,529 INFO HandlerThread:2100 [handler.py:handle_request_defer():184] handle defer: 0 +2024-05-23 13:06:21,529 DEBUG SenderThread:2100 [sender.py:send_request():405] send_request: defer +2024-05-23 13:06:21,529 INFO SenderThread:2100 [sender.py:send_request_defer():609] handle sender defer: 0 +2024-05-23 13:06:21,529 INFO SenderThread:2100 [sender.py:transition_state():613] send defer: 1 +2024-05-23 13:06:21,529 DEBUG HandlerThread:2100 [handler.py:handle_request():158] handle_request: defer +2024-05-23 13:06:21,529 INFO HandlerThread:2100 [handler.py:handle_request_defer():184] handle defer: 1 +2024-05-23 13:06:21,529 DEBUG SenderThread:2100 [sender.py:send_request():405] send_request: defer +2024-05-23 13:06:21,529 INFO SenderThread:2100 [sender.py:send_request_defer():609] handle sender defer: 1 +2024-05-23 13:06:21,529 INFO SenderThread:2100 [sender.py:transition_state():613] send defer: 2 +2024-05-23 13:06:21,529 DEBUG HandlerThread:2100 [handler.py:handle_request():158] handle_request: defer +2024-05-23 13:06:21,529 INFO HandlerThread:2100 [handler.py:handle_request_defer():184] handle defer: 2 +2024-05-23 13:06:21,529 INFO HandlerThread:2100 [system_monitor.py:finish():203] Stopping system monitor +2024-05-23 13:06:21,529 DEBUG SystemMonitor:2100 [system_monitor.py:_start():172] Starting system metrics aggregation loop +2024-05-23 13:06:21,529 DEBUG SystemMonitor:2100 [system_monitor.py:_start():179] Finished system metrics aggregation loop +2024-05-23 13:06:21,530 DEBUG SystemMonitor:2100 [system_monitor.py:_start():183] Publishing last batch of metrics +2024-05-23 13:06:21,530 INFO HandlerThread:2100 [interfaces.py:finish():200] Joined cpu monitor +2024-05-23 13:06:21,531 INFO HandlerThread:2100 [interfaces.py:finish():200] Joined disk monitor +2024-05-23 13:06:21,532 INFO HandlerThread:2100 [interfaces.py:finish():200] Joined memory monitor +2024-05-23 13:06:21,532 INFO HandlerThread:2100 [interfaces.py:finish():200] Joined network monitor +2024-05-23 13:06:21,532 DEBUG SenderThread:2100 [sender.py:send_request():405] send_request: defer +2024-05-23 13:06:21,532 INFO SenderThread:2100 [sender.py:send_request_defer():609] handle sender defer: 2 +2024-05-23 13:06:21,532 INFO SenderThread:2100 [sender.py:transition_state():613] send defer: 3 +2024-05-23 13:06:21,532 DEBUG SenderThread:2100 [sender.py:send():378] send: stats +2024-05-23 13:06:21,533 DEBUG HandlerThread:2100 [handler.py:handle_request():158] handle_request: defer +2024-05-23 13:06:21,533 INFO HandlerThread:2100 [handler.py:handle_request_defer():184] handle defer: 3 +2024-05-23 13:06:21,534 DEBUG SenderThread:2100 [sender.py:send_request():405] send_request: defer +2024-05-23 13:06:21,534 INFO SenderThread:2100 [sender.py:send_request_defer():609] handle sender defer: 3 +2024-05-23 13:06:21,534 INFO SenderThread:2100 [sender.py:transition_state():613] send defer: 4 +2024-05-23 13:06:21,534 DEBUG HandlerThread:2100 [handler.py:handle_request():158] handle_request: defer +2024-05-23 13:06:21,534 INFO HandlerThread:2100 [handler.py:handle_request_defer():184] handle defer: 4 +2024-05-23 13:06:21,534 DEBUG SenderThread:2100 [sender.py:send_request():405] send_request: defer +2024-05-23 13:06:21,534 INFO SenderThread:2100 [sender.py:send_request_defer():609] handle sender defer: 4 +2024-05-23 13:06:21,534 INFO SenderThread:2100 [sender.py:transition_state():613] send defer: 5 +2024-05-23 13:06:21,534 DEBUG HandlerThread:2100 [handler.py:handle_request():158] handle_request: defer +2024-05-23 13:06:21,534 INFO HandlerThread:2100 [handler.py:handle_request_defer():184] handle defer: 5 +2024-05-23 13:06:21,534 DEBUG SenderThread:2100 [sender.py:send():378] send: summary +2024-05-23 13:06:21,535 INFO SenderThread:2100 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end +2024-05-23 13:06:21,535 DEBUG SenderThread:2100 [sender.py:send_request():405] send_request: defer +2024-05-23 13:06:21,535 INFO SenderThread:2100 [sender.py:send_request_defer():609] handle sender defer: 5 +2024-05-23 13:06:21,535 INFO SenderThread:2100 [sender.py:transition_state():613] send defer: 6 +2024-05-23 13:06:21,535 DEBUG HandlerThread:2100 [handler.py:handle_request():158] handle_request: defer +2024-05-23 13:06:21,535 INFO HandlerThread:2100 [handler.py:handle_request_defer():184] handle defer: 6 +2024-05-23 13:06:21,536 DEBUG SenderThread:2100 [sender.py:send_request():405] send_request: defer +2024-05-23 13:06:21,536 INFO SenderThread:2100 [sender.py:send_request_defer():609] handle sender defer: 6 +2024-05-23 13:06:21,540 DEBUG HandlerThread:2100 [handler.py:handle_request():158] handle_request: status_report +2024-05-23 13:06:21,608 INFO SenderThread:2100 [sender.py:transition_state():613] send defer: 7 +2024-05-23 13:06:21,608 DEBUG HandlerThread:2100 [handler.py:handle_request():158] handle_request: defer +2024-05-23 13:06:21,608 INFO HandlerThread:2100 [handler.py:handle_request_defer():184] handle defer: 7 +2024-05-23 13:06:21,608 DEBUG SenderThread:2100 [sender.py:send_request():405] send_request: defer +2024-05-23 13:06:21,609 INFO SenderThread:2100 [sender.py:send_request_defer():609] handle sender defer: 7 +2024-05-23 13:06:22,328 INFO Thread-12 :2100 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240523_130610-knsm0gsf/files/config.yaml +2024-05-23 13:06:22,329 INFO Thread-12 :2100 [dir_watcher.py:_on_file_created():271] file/dir created: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240523_130610-knsm0gsf/files/wandb-summary.json +2024-05-23 13:06:22,526 DEBUG HandlerThread:2100 [handler.py:handle_request():158] handle_request: poll_exit +2024-05-23 13:06:22,895 INFO SenderThread:2100 [sender.py:transition_state():613] send defer: 8 +2024-05-23 13:06:22,895 DEBUG SenderThread:2100 [sender.py:send_request():405] send_request: poll_exit +2024-05-23 13:06:22,895 DEBUG HandlerThread:2100 [handler.py:handle_request():158] handle_request: defer +2024-05-23 13:06:22,895 INFO HandlerThread:2100 [handler.py:handle_request_defer():184] handle defer: 8 +2024-05-23 13:06:22,895 DEBUG SenderThread:2100 [sender.py:send_request():405] send_request: defer +2024-05-23 13:06:22,895 INFO SenderThread:2100 [sender.py:send_request_defer():609] handle sender defer: 8 +2024-05-23 13:06:22,895 INFO SenderThread:2100 [job_builder.py:build():432] Attempting to build job artifact +2024-05-23 13:06:22,896 INFO SenderThread:2100 [job_builder.py:_get_source_type():576] no source found +2024-05-23 13:06:22,896 INFO SenderThread:2100 [sender.py:transition_state():613] send defer: 9 +2024-05-23 13:06:22,896 DEBUG HandlerThread:2100 [handler.py:handle_request():158] handle_request: defer +2024-05-23 13:06:22,896 INFO HandlerThread:2100 [handler.py:handle_request_defer():184] handle defer: 9 +2024-05-23 13:06:22,896 DEBUG SenderThread:2100 [sender.py:send_request():405] send_request: defer +2024-05-23 13:06:22,896 INFO SenderThread:2100 [sender.py:send_request_defer():609] handle sender defer: 9 +2024-05-23 13:06:22,896 INFO SenderThread:2100 [dir_watcher.py:finish():358] shutting down directory watcher +2024-05-23 13:06:23,330 INFO SenderThread:2100 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240523_130610-knsm0gsf/files/output.log +2024-05-23 13:06:23,330 INFO SenderThread:2100 [dir_watcher.py:finish():388] scan: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240523_130610-knsm0gsf/files +2024-05-23 13:06:23,331 INFO SenderThread:2100 [dir_watcher.py:finish():402] scan save: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240523_130610-knsm0gsf/files/output.log output.log +2024-05-23 13:06:23,331 INFO SenderThread:2100 [dir_watcher.py:finish():402] scan save: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240523_130610-knsm0gsf/files/wandb-metadata.json wandb-metadata.json +2024-05-23 13:06:23,333 INFO SenderThread:2100 [dir_watcher.py:finish():402] scan save: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240523_130610-knsm0gsf/files/requirements.txt requirements.txt +2024-05-23 13:06:23,333 INFO SenderThread:2100 [dir_watcher.py:finish():402] scan save: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240523_130610-knsm0gsf/files/config.yaml config.yaml +2024-05-23 13:06:23,333 INFO SenderThread:2100 [dir_watcher.py:finish():402] scan save: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240523_130610-knsm0gsf/files/wandb-summary.json wandb-summary.json +2024-05-23 13:06:23,333 INFO SenderThread:2100 [sender.py:transition_state():613] send defer: 10 +2024-05-23 13:06:23,334 DEBUG HandlerThread:2100 [handler.py:handle_request():158] handle_request: defer +2024-05-23 13:06:23,334 INFO HandlerThread:2100 [handler.py:handle_request_defer():184] handle defer: 10 +2024-05-23 13:06:23,334 DEBUG SenderThread:2100 [sender.py:send_request():405] send_request: defer +2024-05-23 13:06:23,334 INFO SenderThread:2100 [sender.py:send_request_defer():609] handle sender defer: 10 +2024-05-23 13:06:23,334 INFO SenderThread:2100 [file_pusher.py:finish():169] shutting down file pusher +2024-05-23 13:06:23,527 DEBUG HandlerThread:2100 [handler.py:handle_request():158] handle_request: poll_exit +2024-05-23 13:06:23,527 DEBUG SenderThread:2100 [sender.py:send_request():405] send_request: poll_exit +2024-05-23 13:06:23,584 INFO wandb-upload_0:2100 [upload_job.py:push():130] Uploaded file /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240523_130610-knsm0gsf/files/output.log +2024-05-23 13:06:23,943 INFO wandb-upload_1:2100 [upload_job.py:push():130] Uploaded file /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240523_130610-knsm0gsf/files/requirements.txt +2024-05-23 13:06:23,946 INFO wandb-upload_3:2100 [upload_job.py:push():130] Uploaded file /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240523_130610-knsm0gsf/files/wandb-summary.json +2024-05-23 13:06:23,964 INFO wandb-upload_2:2100 [upload_job.py:push():130] Uploaded file /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240523_130610-knsm0gsf/files/config.yaml +2024-05-23 13:06:24,164 INFO Thread-11 (_thread_body):2100 [sender.py:transition_state():613] send defer: 11 +2024-05-23 13:06:24,164 DEBUG HandlerThread:2100 [handler.py:handle_request():158] handle_request: defer +2024-05-23 13:06:24,164 INFO HandlerThread:2100 [handler.py:handle_request_defer():184] handle defer: 11 +2024-05-23 13:06:24,165 DEBUG SenderThread:2100 [sender.py:send_request():405] send_request: defer +2024-05-23 13:06:24,165 INFO SenderThread:2100 [sender.py:send_request_defer():609] handle sender defer: 11 +2024-05-23 13:06:24,165 INFO SenderThread:2100 [file_pusher.py:join():175] waiting for file pusher +2024-05-23 13:06:24,165 INFO SenderThread:2100 [sender.py:transition_state():613] send defer: 12 +2024-05-23 13:06:24,165 DEBUG HandlerThread:2100 [handler.py:handle_request():158] handle_request: defer +2024-05-23 13:06:24,165 INFO HandlerThread:2100 [handler.py:handle_request_defer():184] handle defer: 12 +2024-05-23 13:06:24,165 DEBUG SenderThread:2100 [sender.py:send_request():405] send_request: defer +2024-05-23 13:06:24,165 INFO SenderThread:2100 [sender.py:send_request_defer():609] handle sender defer: 12 +2024-05-23 13:06:24,165 INFO SenderThread:2100 [file_stream.py:finish():601] file stream finish called +2024-05-23 13:06:24,242 INFO SenderThread:2100 [file_stream.py:finish():605] file stream finish is done +2024-05-23 13:06:24,242 INFO SenderThread:2100 [sender.py:transition_state():613] send defer: 13 +2024-05-23 13:06:24,243 DEBUG HandlerThread:2100 [handler.py:handle_request():158] handle_request: defer +2024-05-23 13:06:24,243 INFO HandlerThread:2100 [handler.py:handle_request_defer():184] handle defer: 13 +2024-05-23 13:06:24,243 DEBUG SenderThread:2100 [sender.py:send_request():405] send_request: defer +2024-05-23 13:06:24,243 INFO SenderThread:2100 [sender.py:send_request_defer():609] handle sender defer: 13 +2024-05-23 13:06:24,243 INFO SenderThread:2100 [sender.py:transition_state():613] send defer: 14 +2024-05-23 13:06:24,243 DEBUG HandlerThread:2100 [handler.py:handle_request():158] handle_request: defer +2024-05-23 13:06:24,243 INFO HandlerThread:2100 [handler.py:handle_request_defer():184] handle defer: 14 +2024-05-23 13:06:24,243 DEBUG SenderThread:2100 [sender.py:send():378] send: final +2024-05-23 13:06:24,243 DEBUG SenderThread:2100 [sender.py:send():378] send: footer +2024-05-23 13:06:24,243 DEBUG SenderThread:2100 [sender.py:send_request():405] send_request: defer +2024-05-23 13:06:24,243 INFO SenderThread:2100 [sender.py:send_request_defer():609] handle sender defer: 14 +2024-05-23 13:06:24,244 DEBUG HandlerThread:2100 [handler.py:handle_request():158] handle_request: poll_exit +2024-05-23 13:06:24,244 DEBUG SenderThread:2100 [sender.py:send_request():405] send_request: poll_exit +2024-05-23 13:06:24,244 DEBUG HandlerThread:2100 [handler.py:handle_request():158] handle_request: poll_exit +2024-05-23 13:06:24,244 DEBUG HandlerThread:2100 [handler.py:handle_request():158] handle_request: server_info +2024-05-23 13:06:24,244 DEBUG HandlerThread:2100 [handler.py:handle_request():158] handle_request: get_summary +2024-05-23 13:06:24,244 DEBUG HandlerThread:2100 [handler.py:handle_request():158] handle_request: sampled_history +2024-05-23 13:06:24,244 DEBUG HandlerThread:2100 [handler.py:handle_request():158] handle_request: internal_messages +2024-05-23 13:06:24,245 DEBUG SenderThread:2100 [sender.py:send_request():405] send_request: poll_exit +2024-05-23 13:06:24,245 DEBUG SenderThread:2100 [sender.py:send_request():405] send_request: server_info +2024-05-23 13:06:24,297 INFO MainThread:2100 [wandb_run.py:_footer_history_summary_info():3994] rendering history +2024-05-23 13:06:24,297 INFO MainThread:2100 [wandb_run.py:_footer_history_summary_info():4026] rendering summary +2024-05-23 13:06:24,297 INFO MainThread:2100 [wandb_run.py:_footer_sync_info():3953] logging synced files +2024-05-23 13:06:24,298 DEBUG HandlerThread:2100 [handler.py:handle_request():158] handle_request: shutdown +2024-05-23 13:06:24,298 INFO HandlerThread:2100 [handler.py:finish():882] shutting down handler +2024-05-23 13:06:25,245 INFO WriterThread:2100 [datastore.py:close():296] close: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240523_130610-knsm0gsf/run-knsm0gsf.wandb +2024-05-23 13:06:25,297 INFO SenderThread:2100 [sender.py:finish():1545] shutting down sender +2024-05-23 13:06:25,297 INFO SenderThread:2100 [file_pusher.py:finish():169] shutting down file pusher +2024-05-23 13:06:25,297 INFO SenderThread:2100 [file_pusher.py:join():175] waiting for file pusher diff --git a/lm-evaluation-harness/wandb/run-20240523_130610-knsm0gsf/logs/debug.log b/lm-evaluation-harness/wandb/run-20240523_130610-knsm0gsf/logs/debug.log new file mode 100644 index 0000000000000000000000000000000000000000..0fb4a02f789eef53598a0e566b4ec411097a1c6e --- /dev/null +++ b/lm-evaluation-harness/wandb/run-20240523_130610-knsm0gsf/logs/debug.log @@ -0,0 +1,29 @@ +2024-05-23 13:06:10,041 INFO MainThread:1945 [wandb_setup.py:_flush():76] Current SDK version is 0.17.0 +2024-05-23 13:06:10,041 INFO MainThread:1945 [wandb_setup.py:_flush():76] Configure stats pid to 1945 +2024-05-23 13:06:10,041 INFO MainThread:1945 [wandb_setup.py:_flush():76] Loading settings from /root/.config/wandb/settings +2024-05-23 13:06:10,041 INFO MainThread:1945 [wandb_setup.py:_flush():76] Loading settings from /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/settings +2024-05-23 13:06:10,041 INFO MainThread:1945 [wandb_setup.py:_flush():76] Loading settings from environment variables: {} +2024-05-23 13:06:10,041 INFO MainThread:1945 [wandb_setup.py:_flush():76] Applying setup settings: {'_disable_service': False} +2024-05-23 13:06:10,041 WARNING MainThread:1945 [wandb_setup.py:_flush():76] Could not find program at -m lm_eval.__main__ +2024-05-23 13:06:10,041 INFO MainThread:1945 [wandb_setup.py:_flush():76] Inferring run settings from compute environment: {'program_relpath': None, 'program': '-m lm_eval.__main__'} +2024-05-23 13:06:10,041 INFO MainThread:1945 [wandb_setup.py:_flush():76] Applying login settings: {} +2024-05-23 13:06:10,041 INFO MainThread:1945 [wandb_init.py:_log_setup():520] Logging user logs to /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240523_130610-knsm0gsf/logs/debug.log +2024-05-23 13:06:10,041 INFO MainThread:1945 [wandb_init.py:_log_setup():521] Logging internal logs to /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240523_130610-knsm0gsf/logs/debug-internal.log +2024-05-23 13:06:10,041 INFO MainThread:1945 [wandb_init.py:init():560] calling init triggers +2024-05-23 13:06:10,041 INFO MainThread:1945 [wandb_init.py:init():567] wandb.init called with sweep_config: {} +config: {} +2024-05-23 13:06:10,041 INFO MainThread:1945 [wandb_init.py:init():610] starting backend +2024-05-23 13:06:10,041 INFO MainThread:1945 [wandb_init.py:init():614] setting up manager +2024-05-23 13:06:10,044 INFO MainThread:1945 [backend.py:_multiprocessing_setup():105] multiprocessing start_methods=fork,spawn,forkserver, using: spawn +2024-05-23 13:06:10,045 INFO MainThread:1945 [wandb_init.py:init():622] backend started and connected +2024-05-23 13:06:10,049 INFO MainThread:1945 [wandb_init.py:init():711] updated telemetry +2024-05-23 13:06:10,058 INFO MainThread:1945 [wandb_init.py:init():744] communicating run to backend with 90.0 second timeout +2024-05-23 13:06:10,316 INFO MainThread:1945 [wandb_run.py:_on_init():2396] communicating current version +2024-05-23 13:06:10,426 INFO MainThread:1945 [wandb_run.py:_on_init():2405] got version response +2024-05-23 13:06:10,426 INFO MainThread:1945 [wandb_init.py:init():795] starting run threads in backend +2024-05-23 13:06:10,737 INFO MainThread:1945 [wandb_run.py:_console_start():2374] atexit reg +2024-05-23 13:06:10,738 INFO MainThread:1945 [wandb_run.py:_redirect():2229] redirect: wrap_raw +2024-05-23 13:06:10,738 INFO MainThread:1945 [wandb_run.py:_redirect():2294] Wrapping output streams. +2024-05-23 13:06:10,738 INFO MainThread:1945 [wandb_run.py:_redirect():2319] Redirects installed. +2024-05-23 13:06:10,739 INFO MainThread:1945 [wandb_init.py:init():838] run started, returning control to user process +2024-05-23 13:06:25,299 WARNING MsgRouterThr:1945 [router.py:message_loop():77] message_loop has been closed diff --git a/lm-evaluation-harness/wandb/run-20240523_130610-knsm0gsf/run-knsm0gsf.wandb b/lm-evaluation-harness/wandb/run-20240523_130610-knsm0gsf/run-knsm0gsf.wandb new file mode 100644 index 0000000000000000000000000000000000000000..e586f221f7d2a645d5c7a9fa4dfd403a5d2a7d9c Binary files /dev/null and b/lm-evaluation-harness/wandb/run-20240523_130610-knsm0gsf/run-knsm0gsf.wandb differ diff --git a/lm-evaluation-harness/wandb/run-20240523_131208-rb8nf99r/files/config.yaml b/lm-evaluation-harness/wandb/run-20240523_131208-rb8nf99r/files/config.yaml new file mode 100644 index 0000000000000000000000000000000000000000..5e65ed1c1d997b86cc9aa58d1f90a7b5427a8204 --- /dev/null +++ b/lm-evaluation-harness/wandb/run-20240523_131208-rb8nf99r/files/config.yaml @@ -0,0 +1,43 @@ +wandb_version: 1 + +_wandb: + desc: null + value: + python_version: 3.10.12 + cli_version: 0.17.0 + framework: huggingface + huggingface_version: 4.41.1 + is_jupyter_run: false + is_kaggle_kernel: false + start_time: 1716469928 + t: + 1: + - 1 + - 5 + - 11 + - 49 + - 51 + - 53 + - 55 + - 71 + - 98 + - 100 + 2: + - 1 + - 5 + - 11 + - 49 + - 51 + - 53 + - 55 + - 71 + - 98 + - 100 + 3: + - 23 + 4: 3.10.12 + 5: 0.17.0 + 6: 4.41.1 + 8: + - 5 + 13: linux-x86_64 diff --git a/lm-evaluation-harness/wandb/run-20240523_131208-rb8nf99r/files/output.log b/lm-evaluation-harness/wandb/run-20240523_131208-rb8nf99r/files/output.log new file mode 100644 index 0000000000000000000000000000000000000000..11f77b9b90dae92177e708c290884f6c445bf519 --- /dev/null +++ b/lm-evaluation-harness/wandb/run-20240523_131208-rb8nf99r/files/output.log @@ -0,0 +1,34 @@ + +2024-05-23:13:12:09,523 INFO [__main__.py:251] Verbosity set to INFO +2024-05-23:13:12:18,003 INFO [__main__.py:335] Selected Tasks: ['arc_easy', 'hellaswag', 'mrpc', 'openbookqa', 'sst2', 'winogrande'] +2024-05-23:13:12:18,004 INFO [evaluator.py:131] Setting random seed to 0 | Setting numpy seed to 1234 | Setting torch manual seed to 1234 +2024-05-23:13:12:18,006 INFO [evaluator.py:177] Initializing hf model, with arguments: {'pretrained': '/mnt/weka/peacock/experiments/llama/checkpoint/llamav2-3b//hf_ckpt//global_step8000'} +2024-05-23:13:12:20,311 INFO [huggingface.py:164] Using device 'cuda' +Traceback (most recent call last): + File "/usr/lib/python3.10/runpy.py", line 196, in _run_module_as_main + return _run_code(code, main_globals, None, + File "/usr/lib/python3.10/runpy.py", line 86, in _run_code + exec(code, run_globals) + File "/mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/lm_eval/__main__.py", line 417, in + cli_evaluate() + File "/mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/lm_eval/__main__.py", line 341, in cli_evaluate + results = evaluator.simple_evaluate( + File "/mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/lm_eval/utils.py", line 288, in _wrapper + return fn(*args, **kwargs) + File "/mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/lm_eval/evaluator.py", line 180, in simple_evaluate + lm = lm_eval.api.registry.get_model(model).create_from_arg_string( + File "/mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/lm_eval/api/model.py", line 134, in create_from_arg_string + return cls(**args, **args2) + File "/mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/lm_eval/models/huggingface.py", line 190, in __init__ + self._get_config( + File "/mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/lm_eval/models/huggingface.py", line 471, in _get_config + self._config = transformers.AutoConfig.from_pretrained( + File "/usr/local/lib/python3.10/dist-packages/transformers/models/auto/configuration_auto.py", line 934, in from_pretrained + config_dict, unused_kwargs = PretrainedConfig.get_config_dict(pretrained_model_name_or_path, **kwargs) + File "/usr/local/lib/python3.10/dist-packages/transformers/configuration_utils.py", line 632, in get_config_dict + config_dict, kwargs = cls._get_config_dict(pretrained_model_name_or_path, **kwargs) + File "/usr/local/lib/python3.10/dist-packages/transformers/configuration_utils.py", line 689, in _get_config_dict + resolved_config_file = cached_file( + File "/usr/local/lib/python3.10/dist-packages/transformers/utils/hub.py", line 370, in cached_file + raise EnvironmentError( +OSError: /mnt/weka/peacock/experiments/llama/checkpoint/llamav2-3b//hf_ckpt//global_step8000 does not appear to have a file named config.json. Checkout 'https://huggingface.co//mnt/weka/peacock/experiments/llama/checkpoint/llamav2-3b//hf_ckpt//global_step8000/tree/main' for available files. \ No newline at end of file diff --git a/lm-evaluation-harness/wandb/run-20240523_131208-rb8nf99r/files/requirements.txt b/lm-evaluation-harness/wandb/run-20240523_131208-rb8nf99r/files/requirements.txt new file mode 100644 index 0000000000000000000000000000000000000000..f675c3016b5332c1acf28f436e0b60adeead9c12 --- /dev/null +++ b/lm-evaluation-harness/wandb/run-20240523_131208-rb8nf99r/files/requirements.txt @@ -0,0 +1,155 @@ +DataProperty==1.0.1 +GitPython==3.1.43 +Jinja2==3.1.4 +Markdown==3.6 +MarkupSafe==2.1.5 +Pillow-SIMD==7.0.0.post3 +PyYAML==6.0 +Werkzeug==3.0.3 +absl-py==2.1.0 +accelerate==0.30.1 +aiohttp==3.9.5 +aiosignal==1.3.1 +async-timeout==4.0.3 +attrs==23.2.0 +av==9.2.0 +cachetools==5.3.3 +certifi==2024.2.2 +cffi==1.15.1 +cfgv==3.4.0 +chardet==5.2.0 +charset-normalizer==3.3.2 +click==8.1.7 +cmake==3.29.2 +colorama==0.4.6 +datasets==2.19.1 +deepspeed==0.12.4+hpu.synapse.v1.15.1 +dill==0.3.8 +distlib==0.3.8 +docker-pycreds==0.4.0 +einops==0.8.0 +evaluate==0.4.2 +exceptiongroup==1.2.1 +expecttest==0.2.1 +filelock==3.14.0 +frozenlist==1.4.1 +fsspec==2024.3.1 +gitdb==4.0.11 +google-auth-oauthlib==0.4.6 +google-auth==2.29.0 +grpcio==1.63.0 +habana-media-loader==1.15.1.15 +habana-pyhlml==1.15.1.15 +habana-torch-dataloader==1.15.1.15 +habana-torch-plugin==1.15.1.15 +habana_gpu_migration==1.15.1.15 +habana_quantization_toolkit==1.15.1.15 +hjson==3.1.0 +huggingface-hub==0.23.1 +identify==2.5.36 +idna==3.7 +iniconfig==2.0.0 +joblib==1.4.2 +jsonlines==4.0.0 +lightning-habana==1.4.0 +lightning-utilities==0.11.2 +lightning==2.2.0.post0 +lm_eval==0.4.2 +lm_eval==0.4.2 +lm_eval==0.4.2 +lxml==5.2.2 +mbstrdecoder==1.1.3 +more-itertools==10.2.0 +mpi4py==3.1.4 +mpmath==1.3.0 +multidict==6.0.5 +multiprocess==0.70.16 +networkx==3.3 +ninja==1.11.1.1 +nltk==3.8.1 +nodeenv==1.8.0 +numexpr==2.10.0 +numpy==1.23.5 +oauthlib==3.2.2 +packaging==24.0 +pandas==2.0.1 +pathspec==0.12.1 +pathvalidate==3.2.0 +peft==0.11.1 +perfetto==0.7.0 +pillow==10.3.0 +pip==22.0.2 +pip==23.3.1 +platformdirs==4.2.1 +pluggy==1.5.0 +portalocker==2.8.2 +pre-commit==3.3.3 +pretty-errors==1.2.25 +protobuf==3.20.3 +psutil==5.9.8 +py-cpuinfo==9.0.0 +pyarrow-hotfix==0.6 +pyarrow==16.1.0 +pyasn1==0.6.0 +pyasn1_modules==0.4.0 +pybind11==2.10.4 +pycparser==2.22 +pydantic==1.10.13 +pynvml==8.0.4 +pytablewriter==1.2.0 +pytest==8.2.0 +python-dateutil==2.9.0.post0 +pytorch-lightning==2.2.4 +pytz==2024.1 +regex==2023.5.5 +requests-oauthlib==2.0.0 +requests==2.31.0 +rouge_score==0.1.2 +rsa==4.9 +sacrebleu==2.4.2 +safetensors==0.4.3 +scikit-learn==1.5.0 +scipy==1.13.1 +sentencepiece==0.2.0 +sentry-sdk==2.3.0 +setproctitle==1.3.3 +setuptools==59.6.0 +setuptools==69.5.1 +six==1.16.0 +smmap==5.0.1 +sqlitedict==2.1.0 +symengine==0.11.0 +sympy==1.12 +tabledata==1.3.3 +tabulate==0.9.0 +tcolorpy==0.1.6 +tdqm==0.0.1 +tensorboard-data-server==0.6.1 +tensorboard-plugin-wit==1.8.1 +tensorboard==2.11.2 +threadpoolctl==3.5.0 +tokenizers==0.19.1 +tomli==2.0.1 +torch==2.2.0a0+git8964477 +torch_tb_profiler==0.4.0 +torchaudio==2.2.0+08901ad +torchdata==0.7.1+5e6f7b7 +torchmetrics==1.4.0 +torchtext==0.17.0+400da5c +torchvision==0.17.0+b2383d4 +tqdm-multiprocess==0.0.11 +tqdm==4.66.4 +transformers==4.41.1 +typepy==1.3.2 +typing_extensions==4.11.0 +tzdata==2024.1 +urllib3==1.26.18 +virtualenv==20.26.1 +wandb==0.17.0 +wheel==0.37.1 +wheel==0.43.0 +word2number==1.1 +xxhash==3.4.1 +yamllint==1.35.1 +yarl==1.9.4 +zstandard==0.22.0 \ No newline at end of file diff --git a/lm-evaluation-harness/wandb/run-20240523_131208-rb8nf99r/files/wandb-metadata.json b/lm-evaluation-harness/wandb/run-20240523_131208-rb8nf99r/files/wandb-metadata.json new file mode 100644 index 0000000000000000000000000000000000000000..8967b1e2cdde4c71df253585b05017b52c99a8e9 --- /dev/null +++ b/lm-evaluation-harness/wandb/run-20240523_131208-rb8nf99r/files/wandb-metadata.json @@ -0,0 +1,850 @@ +{ + "os": "Linux-5.15.0-92-generic-x86_64-with-glibc2.35", + "python": "3.10.12", + "heartbeatAt": "2024-05-23T13:12:09.301225", + "startedAt": "2024-05-23T13:12:08.776728", + "docker": null, + "cuda": null, + "args": [ + "--model", + "hf", + "--model_args", + "pretrained=/mnt/weka/peacock/experiments/llama/checkpoint/llamav2-3b//hf_ckpt//global_step8000", + "--tasks", + "hellaswag,arc_easy,openbookqa,winogrande,sst2,mrpc", + "--batch_size", + "auto", + "--wandb_args", + "project=bharatgpt,group=trial_expt_2" + ], + "state": "running", + "program": "-m lm_eval.__main__", + "codePathLocal": null, + "git": { + "remote": "https://github.com/EleutherAI/lm-evaluation-harness", + "commit": null + }, + "email": null, + "root": "/mnt/weka/peacock/idc/cronscript/lm-evaluation-harness", + "host": "peacock-evaluation-worker-0", + "username": "root", + "executable": "/usr/bin/python3", + "cpu_count": 80, + "cpu_count_logical": 160, + "cpu_freq": { + "current": 2333.7720249999998, + "min": 800.0, + "max": 3400.0 + }, + "cpu_freq_per_core": [ + { + "current": 3388.59, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3388.6, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3318.823, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + } + ], + "disk": { + "/": { + "total": 877.6341285705566, + "used": 211.61694717407227 + } + }, + "memory": { + "total": 1007.4379539489746 + } +} diff --git a/lm-evaluation-harness/wandb/run-20240523_131208-rb8nf99r/files/wandb-summary.json b/lm-evaluation-harness/wandb/run-20240523_131208-rb8nf99r/files/wandb-summary.json new file mode 100644 index 0000000000000000000000000000000000000000..8bf99d152ad35c3699ec8600ecb8b169d4e35875 --- /dev/null +++ b/lm-evaluation-harness/wandb/run-20240523_131208-rb8nf99r/files/wandb-summary.json @@ -0,0 +1 @@ +{"_wandb": {"runtime": 11}} \ No newline at end of file diff --git a/lm-evaluation-harness/wandb/run-20240523_131208-rb8nf99r/logs/debug-internal.log b/lm-evaluation-harness/wandb/run-20240523_131208-rb8nf99r/logs/debug-internal.log new file mode 100644 index 0000000000000000000000000000000000000000..924fb083fcf8175fefbb96ce2d69206e235fe10a --- /dev/null +++ b/lm-evaluation-harness/wandb/run-20240523_131208-rb8nf99r/logs/debug-internal.log @@ -0,0 +1,183 @@ +2024-05-23 13:12:08,797 INFO StreamThr :4625 [internal.py:wandb_internal():85] W&B internal server running at pid: 4625, started at: 2024-05-23 13:12:08.795351 +2024-05-23 13:12:08,802 DEBUG HandlerThread:4625 [handler.py:handle_request():158] handle_request: status +2024-05-23 13:12:08,802 INFO WriterThread:4625 [datastore.py:open_for_write():87] open: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240523_131208-rb8nf99r/run-rb8nf99r.wandb +2024-05-23 13:12:08,805 DEBUG SenderThread:4625 [sender.py:send():378] send: header +2024-05-23 13:12:08,807 DEBUG SenderThread:4625 [sender.py:send():378] send: run +2024-05-23 13:12:09,067 INFO SenderThread:4625 [dir_watcher.py:__init__():211] watching files in: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240523_131208-rb8nf99r/files +2024-05-23 13:12:09,068 INFO SenderThread:4625 [sender.py:_start_run_threads():1123] run started: rb8nf99r with start time 1716469928.795803 +2024-05-23 13:12:09,068 DEBUG HandlerThread:4625 [handler.py:handle_request():158] handle_request: check_version +2024-05-23 13:12:09,068 DEBUG SenderThread:4625 [sender.py:send_request():405] send_request: check_version +2024-05-23 13:12:09,189 DEBUG HandlerThread:4625 [handler.py:handle_request():158] handle_request: run_start +2024-05-23 13:12:09,191 DEBUG HandlerThread:4625 [system_info.py:__init__():26] System info init +2024-05-23 13:12:09,192 DEBUG HandlerThread:4625 [system_info.py:__init__():41] System info init done +2024-05-23 13:12:09,192 INFO HandlerThread:4625 [system_monitor.py:start():194] Starting system monitor +2024-05-23 13:12:09,192 INFO SystemMonitor:4625 [system_monitor.py:_start():158] Starting system asset monitoring threads +2024-05-23 13:12:09,192 INFO HandlerThread:4625 [system_monitor.py:probe():214] Collecting system info +2024-05-23 13:12:09,198 INFO SystemMonitor:4625 [interfaces.py:start():188] Started cpu monitoring +2024-05-23 13:12:09,199 INFO SystemMonitor:4625 [interfaces.py:start():188] Started disk monitoring +2024-05-23 13:12:09,199 INFO SystemMonitor:4625 [interfaces.py:start():188] Started memory monitoring +2024-05-23 13:12:09,201 INFO SystemMonitor:4625 [interfaces.py:start():188] Started network monitoring +2024-05-23 13:12:09,301 DEBUG HandlerThread:4625 [system_info.py:probe():150] Probing system +2024-05-23 13:12:09,304 DEBUG HandlerThread:4625 [system_info.py:_probe_git():135] Probing git +2024-05-23 13:12:09,314 ERROR HandlerThread:4625 [gitlib.py:root():92] git root error: Cmd('git') failed due to: exit code(128) + cmdline: git rev-parse --show-toplevel + stderr: 'fatal: detected dubious ownership in repository at '/mnt/weka/peacock/idc/cronscript/lm-evaluation-harness' +To add an exception for this directory, call: + + git config --global --add safe.directory /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness' +2024-05-23 13:12:09,314 DEBUG HandlerThread:4625 [system_info.py:_probe_git():143] Probing git done +2024-05-23 13:12:09,314 DEBUG HandlerThread:4625 [system_info.py:probe():198] Probing system done +2024-05-23 13:12:09,314 DEBUG HandlerThread:4625 [system_monitor.py:probe():223] {'os': 'Linux-5.15.0-92-generic-x86_64-with-glibc2.35', 'python': '3.10.12', 'heartbeatAt': '2024-05-23T13:12:09.301225', 'startedAt': '2024-05-23T13:12:08.776728', 'docker': None, 'cuda': None, 'args': ('--model', 'hf', '--model_args', 'pretrained=/mnt/weka/peacock/experiments/llama/checkpoint/llamav2-3b//hf_ckpt//global_step8000', '--tasks', 'hellaswag,arc_easy,openbookqa,winogrande,sst2,mrpc', '--batch_size', 'auto', '--wandb_args', 'project=bharatgpt,group=trial_expt_2'), 'state': 'running', 'program': '-m lm_eval.__main__', 'codePathLocal': None, 'git': {'remote': 'https://github.com/EleutherAI/lm-evaluation-harness', 'commit': None}, 'email': None, 'root': '/mnt/weka/peacock/idc/cronscript/lm-evaluation-harness', 'host': 'peacock-evaluation-worker-0', 'username': 'root', 'executable': '/usr/bin/python3', 'cpu_count': 80, 'cpu_count_logical': 160, 'cpu_freq': {'current': 2333.7720249999998, 'min': 800.0, 'max': 3400.0}, 'cpu_freq_per_core': [{'current': 3388.59, 'min': 800.0, 'max': 3400.0}, {'current': 3388.6, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 3318.823, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}], 'disk': {'/': {'total': 877.6341285705566, 'used': 211.61694717407227}}, 'memory': {'total': 1007.4379539489746}} +2024-05-23 13:12:09,315 INFO HandlerThread:4625 [system_monitor.py:probe():224] Finished collecting system info +2024-05-23 13:12:09,315 INFO HandlerThread:4625 [system_monitor.py:probe():227] Publishing system info +2024-05-23 13:12:09,318 INFO HandlerThread:4625 [system_monitor.py:probe():229] Finished publishing system info +2024-05-23 13:12:09,323 DEBUG SenderThread:4625 [sender.py:send():378] send: files +2024-05-23 13:12:09,323 INFO SenderThread:4625 [sender.py:_save_file():1389] saving file wandb-metadata.json with policy now +2024-05-23 13:12:09,505 DEBUG HandlerThread:4625 [handler.py:handle_request():158] handle_request: python_packages +2024-05-23 13:12:09,505 DEBUG SenderThread:4625 [sender.py:send_request():405] send_request: python_packages +2024-05-23 13:12:09,507 DEBUG HandlerThread:4625 [handler.py:handle_request():158] handle_request: stop_status +2024-05-23 13:12:09,507 DEBUG SenderThread:4625 [sender.py:send_request():405] send_request: stop_status +2024-05-23 13:12:09,636 DEBUG SenderThread:4625 [sender.py:send():378] send: telemetry +2024-05-23 13:12:09,886 INFO wandb-upload_0:4625 [upload_job.py:push():130] Uploaded file /tmp/tmpvv3rbkfuwandb/oqqlte9f-wandb-metadata.json +2024-05-23 13:12:10,069 INFO Thread-12 :4625 [dir_watcher.py:_on_file_created():271] file/dir created: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240523_131208-rb8nf99r/files/requirements.txt +2024-05-23 13:12:10,069 INFO Thread-12 :4625 [dir_watcher.py:_on_file_created():271] file/dir created: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240523_131208-rb8nf99r/files/output.log +2024-05-23 13:12:10,069 INFO Thread-12 :4625 [dir_watcher.py:_on_file_created():271] file/dir created: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240523_131208-rb8nf99r/files/wandb-metadata.json +2024-05-23 13:12:12,069 INFO Thread-12 :4625 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240523_131208-rb8nf99r/files/output.log +2024-05-23 13:12:14,641 DEBUG HandlerThread:4625 [handler.py:handle_request():158] handle_request: status_report +2024-05-23 13:12:20,007 DEBUG HandlerThread:4625 [handler.py:handle_request():158] handle_request: status_report +2024-05-23 13:12:20,076 INFO Thread-12 :4625 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240523_131208-rb8nf99r/files/output.log +2024-05-23 13:12:20,323 DEBUG SenderThread:4625 [sender.py:send():378] send: exit +2024-05-23 13:12:20,323 INFO SenderThread:4625 [sender.py:send_exit():585] handling exit code: 1 +2024-05-23 13:12:20,323 INFO SenderThread:4625 [sender.py:send_exit():587] handling runtime: 11 +2024-05-23 13:12:20,325 INFO SenderThread:4625 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end +2024-05-23 13:12:20,325 INFO SenderThread:4625 [sender.py:send_exit():593] send defer +2024-05-23 13:12:20,325 DEBUG HandlerThread:4625 [handler.py:handle_request():158] handle_request: defer +2024-05-23 13:12:20,325 INFO HandlerThread:4625 [handler.py:handle_request_defer():184] handle defer: 0 +2024-05-23 13:12:20,325 DEBUG SenderThread:4625 [sender.py:send_request():405] send_request: defer +2024-05-23 13:12:20,325 INFO SenderThread:4625 [sender.py:send_request_defer():609] handle sender defer: 0 +2024-05-23 13:12:20,325 INFO SenderThread:4625 [sender.py:transition_state():613] send defer: 1 +2024-05-23 13:12:20,325 DEBUG HandlerThread:4625 [handler.py:handle_request():158] handle_request: defer +2024-05-23 13:12:20,325 INFO HandlerThread:4625 [handler.py:handle_request_defer():184] handle defer: 1 +2024-05-23 13:12:20,325 DEBUG SenderThread:4625 [sender.py:send_request():405] send_request: defer +2024-05-23 13:12:20,325 INFO SenderThread:4625 [sender.py:send_request_defer():609] handle sender defer: 1 +2024-05-23 13:12:20,326 INFO SenderThread:4625 [sender.py:transition_state():613] send defer: 2 +2024-05-23 13:12:20,326 DEBUG HandlerThread:4625 [handler.py:handle_request():158] handle_request: defer +2024-05-23 13:12:20,326 INFO HandlerThread:4625 [handler.py:handle_request_defer():184] handle defer: 2 +2024-05-23 13:12:20,326 INFO HandlerThread:4625 [system_monitor.py:finish():203] Stopping system monitor +2024-05-23 13:12:20,326 DEBUG SystemMonitor:4625 [system_monitor.py:_start():172] Starting system metrics aggregation loop +2024-05-23 13:12:20,326 DEBUG SystemMonitor:4625 [system_monitor.py:_start():179] Finished system metrics aggregation loop +2024-05-23 13:12:20,326 DEBUG SystemMonitor:4625 [system_monitor.py:_start():183] Publishing last batch of metrics +2024-05-23 13:12:20,326 INFO HandlerThread:4625 [interfaces.py:finish():200] Joined cpu monitor +2024-05-23 13:12:20,328 INFO HandlerThread:4625 [interfaces.py:finish():200] Joined disk monitor +2024-05-23 13:12:20,328 INFO HandlerThread:4625 [interfaces.py:finish():200] Joined memory monitor +2024-05-23 13:12:20,328 INFO HandlerThread:4625 [interfaces.py:finish():200] Joined network monitor +2024-05-23 13:12:20,328 DEBUG SenderThread:4625 [sender.py:send_request():405] send_request: defer +2024-05-23 13:12:20,329 INFO SenderThread:4625 [sender.py:send_request_defer():609] handle sender defer: 2 +2024-05-23 13:12:20,329 INFO SenderThread:4625 [sender.py:transition_state():613] send defer: 3 +2024-05-23 13:12:20,329 DEBUG SenderThread:4625 [sender.py:send():378] send: stats +2024-05-23 13:12:20,329 DEBUG HandlerThread:4625 [handler.py:handle_request():158] handle_request: defer +2024-05-23 13:12:20,330 INFO HandlerThread:4625 [handler.py:handle_request_defer():184] handle defer: 3 +2024-05-23 13:12:20,330 DEBUG SenderThread:4625 [sender.py:send_request():405] send_request: defer +2024-05-23 13:12:20,330 INFO SenderThread:4625 [sender.py:send_request_defer():609] handle sender defer: 3 +2024-05-23 13:12:20,330 INFO SenderThread:4625 [sender.py:transition_state():613] send defer: 4 +2024-05-23 13:12:20,330 DEBUG HandlerThread:4625 [handler.py:handle_request():158] handle_request: defer +2024-05-23 13:12:20,330 INFO HandlerThread:4625 [handler.py:handle_request_defer():184] handle defer: 4 +2024-05-23 13:12:20,330 DEBUG SenderThread:4625 [sender.py:send_request():405] send_request: defer +2024-05-23 13:12:20,330 INFO SenderThread:4625 [sender.py:send_request_defer():609] handle sender defer: 4 +2024-05-23 13:12:20,330 INFO SenderThread:4625 [sender.py:transition_state():613] send defer: 5 +2024-05-23 13:12:20,330 DEBUG HandlerThread:4625 [handler.py:handle_request():158] handle_request: defer +2024-05-23 13:12:20,330 INFO HandlerThread:4625 [handler.py:handle_request_defer():184] handle defer: 5 +2024-05-23 13:12:20,331 DEBUG SenderThread:4625 [sender.py:send():378] send: summary +2024-05-23 13:12:20,331 INFO SenderThread:4625 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end +2024-05-23 13:12:20,332 DEBUG SenderThread:4625 [sender.py:send_request():405] send_request: defer +2024-05-23 13:12:20,332 INFO SenderThread:4625 [sender.py:send_request_defer():609] handle sender defer: 5 +2024-05-23 13:12:20,332 INFO SenderThread:4625 [sender.py:transition_state():613] send defer: 6 +2024-05-23 13:12:20,332 DEBUG HandlerThread:4625 [handler.py:handle_request():158] handle_request: defer +2024-05-23 13:12:20,332 INFO HandlerThread:4625 [handler.py:handle_request_defer():184] handle defer: 6 +2024-05-23 13:12:20,332 DEBUG SenderThread:4625 [sender.py:send_request():405] send_request: defer +2024-05-23 13:12:20,332 INFO SenderThread:4625 [sender.py:send_request_defer():609] handle sender defer: 6 +2024-05-23 13:12:20,336 DEBUG HandlerThread:4625 [handler.py:handle_request():158] handle_request: status_report +2024-05-23 13:12:20,797 INFO SenderThread:4625 [sender.py:transition_state():613] send defer: 7 +2024-05-23 13:12:20,797 DEBUG HandlerThread:4625 [handler.py:handle_request():158] handle_request: defer +2024-05-23 13:12:20,797 INFO HandlerThread:4625 [handler.py:handle_request_defer():184] handle defer: 7 +2024-05-23 13:12:20,798 DEBUG SenderThread:4625 [sender.py:send_request():405] send_request: defer +2024-05-23 13:12:20,798 INFO SenderThread:4625 [sender.py:send_request_defer():609] handle sender defer: 7 +2024-05-23 13:12:21,077 INFO Thread-12 :4625 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240523_131208-rb8nf99r/files/config.yaml +2024-05-23 13:12:21,077 INFO Thread-12 :4625 [dir_watcher.py:_on_file_created():271] file/dir created: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240523_131208-rb8nf99r/files/wandb-summary.json +2024-05-23 13:12:21,323 DEBUG HandlerThread:4625 [handler.py:handle_request():158] handle_request: poll_exit +2024-05-23 13:12:21,661 INFO SenderThread:4625 [sender.py:transition_state():613] send defer: 8 +2024-05-23 13:12:21,662 DEBUG SenderThread:4625 [sender.py:send_request():405] send_request: poll_exit +2024-05-23 13:12:21,662 DEBUG HandlerThread:4625 [handler.py:handle_request():158] handle_request: defer +2024-05-23 13:12:21,662 INFO HandlerThread:4625 [handler.py:handle_request_defer():184] handle defer: 8 +2024-05-23 13:12:21,662 DEBUG SenderThread:4625 [sender.py:send_request():405] send_request: defer +2024-05-23 13:12:21,662 INFO SenderThread:4625 [sender.py:send_request_defer():609] handle sender defer: 8 +2024-05-23 13:12:21,662 INFO SenderThread:4625 [job_builder.py:build():432] Attempting to build job artifact +2024-05-23 13:12:21,662 INFO SenderThread:4625 [job_builder.py:_get_source_type():576] no source found +2024-05-23 13:12:21,663 INFO SenderThread:4625 [sender.py:transition_state():613] send defer: 9 +2024-05-23 13:12:21,663 DEBUG HandlerThread:4625 [handler.py:handle_request():158] handle_request: defer +2024-05-23 13:12:21,663 INFO HandlerThread:4625 [handler.py:handle_request_defer():184] handle defer: 9 +2024-05-23 13:12:21,663 DEBUG SenderThread:4625 [sender.py:send_request():405] send_request: defer +2024-05-23 13:12:21,663 INFO SenderThread:4625 [sender.py:send_request_defer():609] handle sender defer: 9 +2024-05-23 13:12:21,663 INFO SenderThread:4625 [dir_watcher.py:finish():358] shutting down directory watcher +2024-05-23 13:12:22,078 INFO SenderThread:4625 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240523_131208-rb8nf99r/files/output.log +2024-05-23 13:12:22,079 INFO SenderThread:4625 [dir_watcher.py:finish():388] scan: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240523_131208-rb8nf99r/files +2024-05-23 13:12:22,079 INFO SenderThread:4625 [dir_watcher.py:finish():402] scan save: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240523_131208-rb8nf99r/files/output.log output.log +2024-05-23 13:12:22,079 INFO SenderThread:4625 [dir_watcher.py:finish():402] scan save: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240523_131208-rb8nf99r/files/requirements.txt requirements.txt +2024-05-23 13:12:22,081 INFO SenderThread:4625 [dir_watcher.py:finish():402] scan save: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240523_131208-rb8nf99r/files/config.yaml config.yaml +2024-05-23 13:12:22,082 INFO SenderThread:4625 [dir_watcher.py:finish():402] scan save: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240523_131208-rb8nf99r/files/wandb-metadata.json wandb-metadata.json +2024-05-23 13:12:22,082 INFO SenderThread:4625 [dir_watcher.py:finish():402] scan save: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240523_131208-rb8nf99r/files/wandb-summary.json wandb-summary.json +2024-05-23 13:12:22,082 INFO SenderThread:4625 [sender.py:transition_state():613] send defer: 10 +2024-05-23 13:12:22,082 DEBUG HandlerThread:4625 [handler.py:handle_request():158] handle_request: defer +2024-05-23 13:12:22,082 INFO HandlerThread:4625 [handler.py:handle_request_defer():184] handle defer: 10 +2024-05-23 13:12:22,082 DEBUG SenderThread:4625 [sender.py:send_request():405] send_request: defer +2024-05-23 13:12:22,082 INFO SenderThread:4625 [sender.py:send_request_defer():609] handle sender defer: 10 +2024-05-23 13:12:22,082 INFO SenderThread:4625 [file_pusher.py:finish():169] shutting down file pusher +2024-05-23 13:12:22,323 DEBUG HandlerThread:4625 [handler.py:handle_request():158] handle_request: poll_exit +2024-05-23 13:12:22,323 DEBUG SenderThread:4625 [sender.py:send_request():405] send_request: poll_exit +2024-05-23 13:12:22,325 INFO wandb-upload_0:4625 [upload_job.py:push():130] Uploaded file /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240523_131208-rb8nf99r/files/output.log +2024-05-23 13:12:22,678 INFO wandb-upload_1:4625 [upload_job.py:push():130] Uploaded file /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240523_131208-rb8nf99r/files/requirements.txt +2024-05-23 13:12:22,701 INFO wandb-upload_2:4625 [upload_job.py:push():130] Uploaded file /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240523_131208-rb8nf99r/files/config.yaml +2024-05-23 13:12:22,734 INFO wandb-upload_3:4625 [upload_job.py:push():130] Uploaded file /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240523_131208-rb8nf99r/files/wandb-summary.json +2024-05-23 13:12:22,934 INFO Thread-11 (_thread_body):4625 [sender.py:transition_state():613] send defer: 11 +2024-05-23 13:12:22,934 DEBUG HandlerThread:4625 [handler.py:handle_request():158] handle_request: defer +2024-05-23 13:12:22,934 INFO HandlerThread:4625 [handler.py:handle_request_defer():184] handle defer: 11 +2024-05-23 13:12:22,934 DEBUG SenderThread:4625 [sender.py:send_request():405] send_request: defer +2024-05-23 13:12:22,935 INFO SenderThread:4625 [sender.py:send_request_defer():609] handle sender defer: 11 +2024-05-23 13:12:22,935 INFO SenderThread:4625 [file_pusher.py:join():175] waiting for file pusher +2024-05-23 13:12:22,935 INFO SenderThread:4625 [sender.py:transition_state():613] send defer: 12 +2024-05-23 13:12:22,935 DEBUG HandlerThread:4625 [handler.py:handle_request():158] handle_request: defer +2024-05-23 13:12:22,935 INFO HandlerThread:4625 [handler.py:handle_request_defer():184] handle defer: 12 +2024-05-23 13:12:22,935 DEBUG SenderThread:4625 [sender.py:send_request():405] send_request: defer +2024-05-23 13:12:22,935 INFO SenderThread:4625 [sender.py:send_request_defer():609] handle sender defer: 12 +2024-05-23 13:12:22,935 INFO SenderThread:4625 [file_stream.py:finish():601] file stream finish called +2024-05-23 13:12:23,030 INFO SenderThread:4625 [file_stream.py:finish():605] file stream finish is done +2024-05-23 13:12:23,030 INFO SenderThread:4625 [sender.py:transition_state():613] send defer: 13 +2024-05-23 13:12:23,030 DEBUG HandlerThread:4625 [handler.py:handle_request():158] handle_request: defer +2024-05-23 13:12:23,030 INFO HandlerThread:4625 [handler.py:handle_request_defer():184] handle defer: 13 +2024-05-23 13:12:23,030 DEBUG SenderThread:4625 [sender.py:send_request():405] send_request: defer +2024-05-23 13:12:23,030 INFO SenderThread:4625 [sender.py:send_request_defer():609] handle sender defer: 13 +2024-05-23 13:12:23,030 INFO SenderThread:4625 [sender.py:transition_state():613] send defer: 14 +2024-05-23 13:12:23,030 DEBUG HandlerThread:4625 [handler.py:handle_request():158] handle_request: defer +2024-05-23 13:12:23,030 INFO HandlerThread:4625 [handler.py:handle_request_defer():184] handle defer: 14 +2024-05-23 13:12:23,031 DEBUG SenderThread:4625 [sender.py:send():378] send: final +2024-05-23 13:12:23,031 DEBUG SenderThread:4625 [sender.py:send():378] send: footer +2024-05-23 13:12:23,031 DEBUG SenderThread:4625 [sender.py:send_request():405] send_request: defer +2024-05-23 13:12:23,031 INFO SenderThread:4625 [sender.py:send_request_defer():609] handle sender defer: 14 +2024-05-23 13:12:23,031 DEBUG HandlerThread:4625 [handler.py:handle_request():158] handle_request: poll_exit +2024-05-23 13:12:23,031 DEBUG SenderThread:4625 [sender.py:send_request():405] send_request: poll_exit +2024-05-23 13:12:23,032 DEBUG HandlerThread:4625 [handler.py:handle_request():158] handle_request: poll_exit +2024-05-23 13:12:23,032 DEBUG HandlerThread:4625 [handler.py:handle_request():158] handle_request: server_info +2024-05-23 13:12:23,032 DEBUG HandlerThread:4625 [handler.py:handle_request():158] handle_request: get_summary +2024-05-23 13:12:23,032 DEBUG HandlerThread:4625 [handler.py:handle_request():158] handle_request: sampled_history +2024-05-23 13:12:23,032 DEBUG HandlerThread:4625 [handler.py:handle_request():158] handle_request: internal_messages +2024-05-23 13:12:23,032 DEBUG SenderThread:4625 [sender.py:send_request():405] send_request: poll_exit +2024-05-23 13:12:23,032 DEBUG SenderThread:4625 [sender.py:send_request():405] send_request: server_info +2024-05-23 13:12:23,088 INFO MainThread:4625 [wandb_run.py:_footer_history_summary_info():3994] rendering history +2024-05-23 13:12:23,088 INFO MainThread:4625 [wandb_run.py:_footer_history_summary_info():4026] rendering summary +2024-05-23 13:12:23,088 INFO MainThread:4625 [wandb_run.py:_footer_sync_info():3953] logging synced files +2024-05-23 13:12:23,089 DEBUG HandlerThread:4625 [handler.py:handle_request():158] handle_request: shutdown +2024-05-23 13:12:23,089 INFO HandlerThread:4625 [handler.py:finish():882] shutting down handler +2024-05-23 13:12:24,032 INFO WriterThread:4625 [datastore.py:close():296] close: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240523_131208-rb8nf99r/run-rb8nf99r.wandb +2024-05-23 13:12:24,088 INFO SenderThread:4625 [sender.py:finish():1545] shutting down sender +2024-05-23 13:12:24,088 INFO SenderThread:4625 [file_pusher.py:finish():169] shutting down file pusher +2024-05-23 13:12:24,088 INFO SenderThread:4625 [file_pusher.py:join():175] waiting for file pusher diff --git a/lm-evaluation-harness/wandb/run-20240523_131208-rb8nf99r/logs/debug.log b/lm-evaluation-harness/wandb/run-20240523_131208-rb8nf99r/logs/debug.log new file mode 100644 index 0000000000000000000000000000000000000000..4c3c1508fef0ce25680434b87cd2d120e62ed943 --- /dev/null +++ b/lm-evaluation-harness/wandb/run-20240523_131208-rb8nf99r/logs/debug.log @@ -0,0 +1,29 @@ +2024-05-23 13:12:08,790 INFO MainThread:4470 [wandb_setup.py:_flush():76] Current SDK version is 0.17.0 +2024-05-23 13:12:08,790 INFO MainThread:4470 [wandb_setup.py:_flush():76] Configure stats pid to 4470 +2024-05-23 13:12:08,790 INFO MainThread:4470 [wandb_setup.py:_flush():76] Loading settings from /root/.config/wandb/settings +2024-05-23 13:12:08,790 INFO MainThread:4470 [wandb_setup.py:_flush():76] Loading settings from /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/settings +2024-05-23 13:12:08,790 INFO MainThread:4470 [wandb_setup.py:_flush():76] Loading settings from environment variables: {} +2024-05-23 13:12:08,790 INFO MainThread:4470 [wandb_setup.py:_flush():76] Applying setup settings: {'_disable_service': False} +2024-05-23 13:12:08,791 WARNING MainThread:4470 [wandb_setup.py:_flush():76] Could not find program at -m lm_eval.__main__ +2024-05-23 13:12:08,791 INFO MainThread:4470 [wandb_setup.py:_flush():76] Inferring run settings from compute environment: {'program_relpath': None, 'program': '-m lm_eval.__main__'} +2024-05-23 13:12:08,791 INFO MainThread:4470 [wandb_setup.py:_flush():76] Applying login settings: {} +2024-05-23 13:12:08,791 INFO MainThread:4470 [wandb_init.py:_log_setup():520] Logging user logs to /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240523_131208-rb8nf99r/logs/debug.log +2024-05-23 13:12:08,791 INFO MainThread:4470 [wandb_init.py:_log_setup():521] Logging internal logs to /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240523_131208-rb8nf99r/logs/debug-internal.log +2024-05-23 13:12:08,791 INFO MainThread:4470 [wandb_init.py:init():560] calling init triggers +2024-05-23 13:12:08,791 INFO MainThread:4470 [wandb_init.py:init():567] wandb.init called with sweep_config: {} +config: {} +2024-05-23 13:12:08,791 INFO MainThread:4470 [wandb_init.py:init():610] starting backend +2024-05-23 13:12:08,791 INFO MainThread:4470 [wandb_init.py:init():614] setting up manager +2024-05-23 13:12:08,794 INFO MainThread:4470 [backend.py:_multiprocessing_setup():105] multiprocessing start_methods=fork,spawn,forkserver, using: spawn +2024-05-23 13:12:08,795 INFO MainThread:4470 [wandb_init.py:init():622] backend started and connected +2024-05-23 13:12:08,798 INFO MainThread:4470 [wandb_init.py:init():711] updated telemetry +2024-05-23 13:12:08,806 INFO MainThread:4470 [wandb_init.py:init():744] communicating run to backend with 90.0 second timeout +2024-05-23 13:12:09,068 INFO MainThread:4470 [wandb_run.py:_on_init():2396] communicating current version +2024-05-23 13:12:09,183 INFO MainThread:4470 [wandb_run.py:_on_init():2405] got version response +2024-05-23 13:12:09,183 INFO MainThread:4470 [wandb_init.py:init():795] starting run threads in backend +2024-05-23 13:12:09,506 INFO MainThread:4470 [wandb_run.py:_console_start():2374] atexit reg +2024-05-23 13:12:09,506 INFO MainThread:4470 [wandb_run.py:_redirect():2229] redirect: wrap_raw +2024-05-23 13:12:09,506 INFO MainThread:4470 [wandb_run.py:_redirect():2294] Wrapping output streams. +2024-05-23 13:12:09,506 INFO MainThread:4470 [wandb_run.py:_redirect():2319] Redirects installed. +2024-05-23 13:12:09,521 INFO MainThread:4470 [wandb_init.py:init():838] run started, returning control to user process +2024-05-23 13:12:24,090 WARNING MsgRouterThr:4470 [router.py:message_loop():77] message_loop has been closed diff --git a/lm-evaluation-harness/wandb/run-20240523_131208-rb8nf99r/run-rb8nf99r.wandb b/lm-evaluation-harness/wandb/run-20240523_131208-rb8nf99r/run-rb8nf99r.wandb new file mode 100644 index 0000000000000000000000000000000000000000..2df1c0df64e98a167cd4929ef44e5c928fff3773 Binary files /dev/null and b/lm-evaluation-harness/wandb/run-20240523_131208-rb8nf99r/run-rb8nf99r.wandb differ diff --git a/lm-evaluation-harness/wandb/run-20240530_070444-c0riftte/files/config.yaml b/lm-evaluation-harness/wandb/run-20240530_070444-c0riftte/files/config.yaml new file mode 100644 index 0000000000000000000000000000000000000000..297fed682df39824390772ba7e6b31bf64a8d76a --- /dev/null +++ b/lm-evaluation-harness/wandb/run-20240530_070444-c0riftte/files/config.yaml @@ -0,0 +1,44 @@ +wandb_version: 1 + +_wandb: + desc: null + value: + python_version: 3.10.12 + cli_version: 0.17.0 + framework: huggingface + huggingface_version: 4.36.2 + is_jupyter_run: false + is_kaggle_kernel: false + start_time: 1717052684 + t: + 1: + - 1 + - 5 + - 11 + - 49 + - 51 + - 53 + - 55 + - 71 + - 98 + - 100 + 2: + - 1 + - 5 + - 11 + - 49 + - 51 + - 53 + - 55 + - 71 + - 98 + - 100 + 3: + - 13 + - 23 + 4: 3.10.12 + 5: 0.17.0 + 6: 4.36.2 + 8: + - 5 + 13: linux-x86_64 diff --git a/lm-evaluation-harness/wandb/run-20240530_070444-c0riftte/files/output.log b/lm-evaluation-harness/wandb/run-20240530_070444-c0riftte/files/output.log new file mode 100644 index 0000000000000000000000000000000000000000..213416055fb917f464b10048e63bf79fedd560c4 --- /dev/null +++ b/lm-evaluation-harness/wandb/run-20240530_070444-c0riftte/files/output.log @@ -0,0 +1,38 @@ + +2024-05-30:07:04:45,753 INFO [__main__.py:251] Verbosity set to INFO +2024-05-30:07:04:54,928 INFO [__main__.py:335] Selected Tasks: ['arc_easy', 'boolq', 'copa', 'mrpc', 'piqa', 'sst2', 'winogrande'] +2024-05-30:07:04:54,929 INFO [evaluator.py:131] Setting random seed to 0 | Setting numpy seed to 1234 | Setting torch manual seed to 1234 +2024-05-30:07:04:54,929 INFO [evaluator.py:177] Initializing hf model, with arguments: {'pretrained': '/mnt/weka/peacock/experiments/llama/eval/checkpoint-english/llamav2-3b/hf/global_step40000', 'tokenizer': '/mnt/weka/peacock/tokenization/trained-tokenizer/enhiben_50k/'} +2024-05-30:07:04:57,213 INFO [huggingface.py:164] Using device 'cuda' +/usr/local/lib/python3.10/dist-packages/habana_frameworks/torch/gpu_migration/torch/cuda/memory.py:36: UserWarning: No need to call empty_cache on HPU. It manages the memory internally in an effcient way. + warnings.warn( +Traceback (most recent call last): + File "/usr/lib/python3.10/runpy.py", line 196, in _run_module_as_main + return _run_code(code, main_globals, None, + File "/usr/lib/python3.10/runpy.py", line 86, in _run_code + exec(code, run_globals) + File "/mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/lm_eval/__main__.py", line 417, in + cli_evaluate() + File "/mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/lm_eval/__main__.py", line 341, in cli_evaluate + results = evaluator.simple_evaluate( + File "/mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/lm_eval/utils.py", line 288, in _wrapper + return fn(*args, **kwargs) + File "/mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/lm_eval/evaluator.py", line 180, in simple_evaluate + lm = lm_eval.api.registry.get_model(model).create_from_arg_string( + File "/mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/lm_eval/api/model.py", line 134, in create_from_arg_string + return cls(**args, **args2) + File "/mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/lm_eval/models/huggingface.py", line 237, in __init__ + self._create_tokenizer( + File "/mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/lm_eval/models/huggingface.py", line 630, in _create_tokenizer + self.tokenizer = transformers.AutoTokenizer.from_pretrained( + File "/usr/local/lib/python3.10/dist-packages/transformers/models/auto/tokenization_auto.py", line 752, in from_pretrained + config = AutoConfig.from_pretrained( + File "/usr/local/lib/python3.10/dist-packages/transformers/models/auto/configuration_auto.py", line 1082, in from_pretrained + config_dict, unused_kwargs = PretrainedConfig.get_config_dict(pretrained_model_name_or_path, **kwargs) + File "/usr/local/lib/python3.10/dist-packages/transformers/configuration_utils.py", line 644, in get_config_dict + config_dict, kwargs = cls._get_config_dict(pretrained_model_name_or_path, **kwargs) + File "/usr/local/lib/python3.10/dist-packages/transformers/configuration_utils.py", line 699, in _get_config_dict + resolved_config_file = cached_file( + File "/usr/local/lib/python3.10/dist-packages/transformers/utils/hub.py", line 360, in cached_file + raise EnvironmentError( +OSError: /mnt/weka/peacock/tokenization/trained-tokenizer/enhiben_50k/ does not appear to have a file named config.json. Checkout 'https://huggingface.co//mnt/weka/peacock/tokenization/trained-tokenizer/enhiben_50k//main' for available files. \ No newline at end of file diff --git a/lm-evaluation-harness/wandb/run-20240530_070444-c0riftte/files/requirements.txt b/lm-evaluation-harness/wandb/run-20240530_070444-c0riftte/files/requirements.txt new file mode 100644 index 0000000000000000000000000000000000000000..7391d0fb5302364497bd6017486c36f0caae0613 --- /dev/null +++ b/lm-evaluation-harness/wandb/run-20240530_070444-c0riftte/files/requirements.txt @@ -0,0 +1,154 @@ +DataProperty==1.0.1 +GitPython==3.1.43 +Jinja2==3.1.4 +Markdown==3.6 +MarkupSafe==2.1.5 +Pillow-SIMD==7.0.0.post3 +PyYAML==6.0 +Werkzeug==3.0.3 +absl-py==2.1.0 +accelerate==0.30.1 +aiohttp==3.9.5 +aiosignal==1.3.1 +async-timeout==4.0.3 +attrs==23.2.0 +av==9.2.0 +cachetools==5.3.3 +certifi==2024.2.2 +cffi==1.15.1 +cfgv==3.4.0 +chardet==5.2.0 +charset-normalizer==3.3.2 +click==8.1.7 +cmake==3.29.2 +colorama==0.4.6 +datasets==2.19.1 +deepspeed==0.12.4+hpu.synapse.v1.15.1 +dill==0.3.8 +distlib==0.3.8 +docker-pycreds==0.4.0 +einops==0.8.0 +evaluate==0.4.2 +exceptiongroup==1.2.1 +expecttest==0.2.1 +filelock==3.14.0 +frozenlist==1.4.1 +fsspec==2024.3.1 +gitdb==4.0.11 +google-auth-oauthlib==0.4.6 +google-auth==2.29.0 +grpcio==1.63.0 +habana-media-loader==1.15.1.15 +habana-pyhlml==1.15.1.15 +habana-torch-dataloader==1.15.1.15 +habana-torch-plugin==1.15.1.15 +habana_gpu_migration==1.15.1.15 +habana_quantization_toolkit==1.15.1.15 +hjson==3.1.0 +huggingface-hub==0.23.2 +identify==2.5.36 +idna==3.7 +iniconfig==2.0.0 +joblib==1.4.2 +jsonlines==4.0.0 +lightning-habana==1.4.0 +lightning-utilities==0.11.2 +lightning==2.2.0.post0 +lm_eval==0.4.2 +lm_eval==0.4.2 +lm_eval==0.4.2 +lxml==5.2.2 +mbstrdecoder==1.1.3 +more-itertools==10.2.0 +mpi4py==3.1.4 +mpmath==1.3.0 +multidict==6.0.5 +multiprocess==0.70.16 +networkx==3.3 +ninja==1.11.1.1 +nltk==3.8.1 +nodeenv==1.8.0 +numexpr==2.10.0 +numpy==1.23.5 +oauthlib==3.2.2 +packaging==24.0 +pandas==2.0.1 +pathspec==0.12.1 +pathvalidate==3.2.0 +peft==0.11.1 +perfetto==0.7.0 +pip==22.0.2 +pip==23.3.1 +platformdirs==4.2.1 +pluggy==1.5.0 +portalocker==2.8.2 +pre-commit==3.3.3 +pretty-errors==1.2.25 +protobuf==3.20.3 +psutil==5.9.8 +py-cpuinfo==9.0.0 +pyarrow-hotfix==0.6 +pyarrow==16.1.0 +pyasn1==0.6.0 +pyasn1_modules==0.4.0 +pybind11==2.10.4 +pycparser==2.22 +pydantic==1.10.13 +pynvml==8.0.4 +pytablewriter==1.2.0 +pytest==8.2.0 +python-dateutil==2.9.0.post0 +pytorch-lightning==2.2.4 +pytz==2024.1 +regex==2023.5.5 +requests-oauthlib==2.0.0 +requests==2.31.0 +rouge_score==0.1.2 +rsa==4.9 +sacrebleu==2.4.2 +safetensors==0.4.3 +scikit-learn==1.5.0 +scipy==1.13.1 +sentencepiece==0.2.0 +sentry-sdk==2.3.1 +setproctitle==1.3.3 +setuptools==59.6.0 +setuptools==69.5.1 +six==1.16.0 +smmap==5.0.1 +sqlitedict==2.1.0 +symengine==0.11.0 +sympy==1.12 +tabledata==1.3.3 +tabulate==0.9.0 +tcolorpy==0.1.6 +tdqm==0.0.1 +tensorboard-data-server==0.6.1 +tensorboard-plugin-wit==1.8.1 +tensorboard==2.11.2 +threadpoolctl==3.5.0 +tokenizers==0.15.2 +tomli==2.0.1 +torch==2.2.0a0+git8964477 +torch_tb_profiler==0.4.0 +torchaudio==2.2.0+08901ad +torchdata==0.7.1+5e6f7b7 +torchmetrics==1.4.0 +torchtext==0.17.0+400da5c +torchvision==0.17.0+b2383d4 +tqdm-multiprocess==0.0.11 +tqdm==4.66.4 +transformers==4.36.2 +typepy==1.3.2 +typing_extensions==4.11.0 +tzdata==2024.1 +urllib3==1.26.18 +virtualenv==20.26.1 +wandb==0.17.0 +wheel==0.37.1 +wheel==0.43.0 +word2number==1.1 +xxhash==3.4.1 +yamllint==1.35.1 +yarl==1.9.4 +zstandard==0.22.0 \ No newline at end of file diff --git a/lm-evaluation-harness/wandb/run-20240530_070444-c0riftte/files/wandb-metadata.json b/lm-evaluation-harness/wandb/run-20240530_070444-c0riftte/files/wandb-metadata.json new file mode 100644 index 0000000000000000000000000000000000000000..34101ad19c05117f28b53fc74d6a9894433225b1 --- /dev/null +++ b/lm-evaluation-harness/wandb/run-20240530_070444-c0riftte/files/wandb-metadata.json @@ -0,0 +1,850 @@ +{ + "os": "Linux-5.15.0-92-generic-x86_64-with-glibc2.35", + "python": "3.10.12", + "heartbeatAt": "2024-05-30T07:04:45.526089", + "startedAt": "2024-05-30T07:04:44.743346", + "docker": null, + "cuda": null, + "args": [ + "--model", + "hf", + "--model_args", + "pretrained=/mnt/weka/peacock/experiments/llama/eval/checkpoint-english/llamav2-3b/hf/global_step40000,tokenizer=/mnt/weka/peacock/tokenization/trained-tokenizer/enhiben_50k/", + "--tasks", + "winogrande,sst2,mrpc,arc_easy,copa,piqa,boolq", + "--batch_size", + "auto", + "--wandb_args", + "project=english-eval,group=exp2,name=global_step40000" + ], + "state": "running", + "program": "-m lm_eval.__main__", + "codePathLocal": null, + "git": { + "remote": "https://github.com/EleutherAI/lm-evaluation-harness", + "commit": null + }, + "email": null, + "root": "/mnt/weka/peacock/idc/cronscript/lm-evaluation-harness", + "host": "peacock-evaluation-debug-20-worker-0", + "username": "root", + "executable": "/usr/bin/python3", + "cpu_count": 80, + "cpu_count_logical": 160, + "cpu_freq": { + "current": 2332.0598375, + "min": 800.0, + "max": 3400.0 + }, + "cpu_freq_per_core": [ + { + "current": 3399.997, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3399.997, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3315.791, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + } + ], + "disk": { + "/": { + "total": 877.6341285705566, + "used": 211.98115158081055 + } + }, + "memory": { + "total": 1007.4605484008789 + } +} diff --git a/lm-evaluation-harness/wandb/run-20240530_070444-c0riftte/files/wandb-summary.json b/lm-evaluation-harness/wandb/run-20240530_070444-c0riftte/files/wandb-summary.json new file mode 100644 index 0000000000000000000000000000000000000000..e4da61201766f19c52f50d698a897d6d33c28936 --- /dev/null +++ b/lm-evaluation-harness/wandb/run-20240530_070444-c0riftte/files/wandb-summary.json @@ -0,0 +1 @@ +{"_wandb": {"runtime": 39}} \ No newline at end of file diff --git a/lm-evaluation-harness/wandb/run-20240530_070444-c0riftte/logs/debug-internal.log b/lm-evaluation-harness/wandb/run-20240530_070444-c0riftte/logs/debug-internal.log new file mode 100644 index 0000000000000000000000000000000000000000..8590c49ebc90810cd1128b8623660d0d0b185928 --- /dev/null +++ b/lm-evaluation-harness/wandb/run-20240530_070444-c0riftte/logs/debug-internal.log @@ -0,0 +1,194 @@ +2024-05-30 07:04:44,765 INFO StreamThr :900 [internal.py:wandb_internal():85] W&B internal server running at pid: 900, started at: 2024-05-30 07:04:44.763108 +2024-05-30 07:04:44,769 DEBUG HandlerThread:900 [handler.py:handle_request():158] handle_request: status +2024-05-30 07:04:44,769 INFO WriterThread:900 [datastore.py:open_for_write():87] open: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240530_070444-c0riftte/run-c0riftte.wandb +2024-05-30 07:04:44,771 DEBUG SenderThread:900 [sender.py:send():378] send: header +2024-05-30 07:04:44,775 DEBUG SenderThread:900 [sender.py:send():378] send: run +2024-05-30 07:04:45,268 INFO SenderThread:900 [dir_watcher.py:__init__():211] watching files in: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240530_070444-c0riftte/files +2024-05-30 07:04:45,268 INFO SenderThread:900 [sender.py:_start_run_threads():1123] run started: c0riftte with start time 1717052684.762957 +2024-05-30 07:04:45,273 DEBUG HandlerThread:900 [handler.py:handle_request():158] handle_request: check_version +2024-05-30 07:04:45,273 DEBUG SenderThread:900 [sender.py:send_request():405] send_request: check_version +2024-05-30 07:04:45,432 DEBUG HandlerThread:900 [handler.py:handle_request():158] handle_request: run_start +2024-05-30 07:04:45,435 DEBUG HandlerThread:900 [system_info.py:__init__():26] System info init +2024-05-30 07:04:45,435 DEBUG HandlerThread:900 [system_info.py:__init__():41] System info init done +2024-05-30 07:04:45,435 INFO HandlerThread:900 [system_monitor.py:start():194] Starting system monitor +2024-05-30 07:04:45,435 INFO SystemMonitor:900 [system_monitor.py:_start():158] Starting system asset monitoring threads +2024-05-30 07:04:45,435 INFO HandlerThread:900 [system_monitor.py:probe():214] Collecting system info +2024-05-30 07:04:45,442 INFO SystemMonitor:900 [interfaces.py:start():188] Started cpu monitoring +2024-05-30 07:04:45,442 INFO SystemMonitor:900 [interfaces.py:start():188] Started disk monitoring +2024-05-30 07:04:45,444 INFO SystemMonitor:900 [interfaces.py:start():188] Started memory monitoring +2024-05-30 07:04:45,444 INFO SystemMonitor:900 [interfaces.py:start():188] Started network monitoring +2024-05-30 07:04:45,526 DEBUG HandlerThread:900 [system_info.py:probe():150] Probing system +2024-05-30 07:04:45,529 DEBUG HandlerThread:900 [system_info.py:_probe_git():135] Probing git +2024-05-30 07:04:45,539 ERROR HandlerThread:900 [gitlib.py:root():92] git root error: Cmd('git') failed due to: exit code(128) + cmdline: git rev-parse --show-toplevel + stderr: 'fatal: detected dubious ownership in repository at '/mnt/weka/peacock/idc/cronscript/lm-evaluation-harness' +To add an exception for this directory, call: + + git config --global --add safe.directory /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness' +2024-05-30 07:04:45,540 DEBUG HandlerThread:900 [system_info.py:_probe_git():143] Probing git done +2024-05-30 07:04:45,540 DEBUG HandlerThread:900 [system_info.py:probe():198] Probing system done +2024-05-30 07:04:45,540 DEBUG HandlerThread:900 [system_monitor.py:probe():223] {'os': 'Linux-5.15.0-92-generic-x86_64-with-glibc2.35', 'python': '3.10.12', 'heartbeatAt': '2024-05-30T07:04:45.526089', 'startedAt': '2024-05-30T07:04:44.743346', 'docker': None, 'cuda': None, 'args': ('--model', 'hf', '--model_args', 'pretrained=/mnt/weka/peacock/experiments/llama/eval/checkpoint-english/llamav2-3b/hf/global_step40000,tokenizer=/mnt/weka/peacock/tokenization/trained-tokenizer/enhiben_50k/', '--tasks', 'winogrande,sst2,mrpc,arc_easy,copa,piqa,boolq', '--batch_size', 'auto', '--wandb_args', 'project=english-eval,group=exp2,name=global_step40000'), 'state': 'running', 'program': '-m lm_eval.__main__', 'codePathLocal': None, 'git': {'remote': 'https://github.com/EleutherAI/lm-evaluation-harness', 'commit': None}, 'email': None, 'root': '/mnt/weka/peacock/idc/cronscript/lm-evaluation-harness', 'host': 'peacock-evaluation-debug-20-worker-0', 'username': 'root', 'executable': '/usr/bin/python3', 'cpu_count': 80, 'cpu_count_logical': 160, 'cpu_freq': {'current': 2332.0598375, 'min': 800.0, 'max': 3400.0}, 'cpu_freq_per_core': [{'current': 3399.997, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 3399.997, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 3315.791, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}], 'disk': {'/': {'total': 877.6341285705566, 'used': 211.98115158081055}}, 'memory': {'total': 1007.4605484008789}} +2024-05-30 07:04:45,540 INFO HandlerThread:900 [system_monitor.py:probe():224] Finished collecting system info +2024-05-30 07:04:45,540 INFO HandlerThread:900 [system_monitor.py:probe():227] Publishing system info +2024-05-30 07:04:45,544 INFO HandlerThread:900 [system_monitor.py:probe():229] Finished publishing system info +2024-05-30 07:04:45,551 DEBUG SenderThread:900 [sender.py:send():378] send: files +2024-05-30 07:04:45,551 INFO SenderThread:900 [sender.py:_save_file():1389] saving file wandb-metadata.json with policy now +2024-05-30 07:04:45,733 DEBUG HandlerThread:900 [handler.py:handle_request():158] handle_request: python_packages +2024-05-30 07:04:45,733 DEBUG SenderThread:900 [sender.py:send_request():405] send_request: python_packages +2024-05-30 07:04:45,734 DEBUG HandlerThread:900 [handler.py:handle_request():158] handle_request: stop_status +2024-05-30 07:04:45,734 DEBUG SenderThread:900 [sender.py:send_request():405] send_request: stop_status +2024-05-30 07:04:45,810 DEBUG SenderThread:900 [sender.py:send():378] send: telemetry +2024-05-30 07:04:46,143 INFO wandb-upload_0:900 [upload_job.py:push():130] Uploaded file /tmp/tmpob4tst5ewandb/wzgzo81w-wandb-metadata.json +2024-05-30 07:04:46,270 INFO Thread-12 :900 [dir_watcher.py:_on_file_created():271] file/dir created: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240530_070444-c0riftte/files/wandb-metadata.json +2024-05-30 07:04:46,271 INFO Thread-12 :900 [dir_watcher.py:_on_file_created():271] file/dir created: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240530_070444-c0riftte/files/requirements.txt +2024-05-30 07:04:46,271 INFO Thread-12 :900 [dir_watcher.py:_on_file_created():271] file/dir created: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240530_070444-c0riftte/files/output.log +2024-05-30 07:04:48,270 INFO Thread-12 :900 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240530_070444-c0riftte/files/output.log +2024-05-30 07:04:49,812 DEBUG HandlerThread:900 [handler.py:handle_request():158] handle_request: status_report +2024-05-30 07:04:54,812 DEBUG HandlerThread:900 [handler.py:handle_request():158] handle_request: status_report +2024-05-30 07:04:56,276 INFO Thread-12 :900 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240530_070444-c0riftte/files/output.log +2024-05-30 07:04:58,292 INFO Thread-12 :900 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240530_070444-c0riftte/files/output.log +2024-05-30 07:05:00,218 DEBUG HandlerThread:900 [handler.py:handle_request():158] handle_request: status_report +2024-05-30 07:05:00,734 DEBUG HandlerThread:900 [handler.py:handle_request():158] handle_request: stop_status +2024-05-30 07:05:00,734 DEBUG SenderThread:900 [sender.py:send_request():405] send_request: stop_status +2024-05-30 07:05:05,876 DEBUG HandlerThread:900 [handler.py:handle_request():158] handle_request: status_report +2024-05-30 07:05:10,877 DEBUG HandlerThread:900 [handler.py:handle_request():158] handle_request: status_report +2024-05-30 07:05:15,734 DEBUG HandlerThread:900 [handler.py:handle_request():158] handle_request: stop_status +2024-05-30 07:05:15,734 DEBUG SenderThread:900 [sender.py:send_request():405] send_request: stop_status +2024-05-30 07:05:16,828 DEBUG HandlerThread:900 [handler.py:handle_request():158] handle_request: status_report +2024-05-30 07:05:17,350 INFO Thread-12 :900 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240530_070444-c0riftte/files/config.yaml +2024-05-30 07:05:22,542 DEBUG HandlerThread:900 [handler.py:handle_request():158] handle_request: status_report +2024-05-30 07:05:24,818 INFO Thread-12 :900 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240530_070444-c0riftte/files/output.log +2024-05-30 07:05:24,836 DEBUG SenderThread:900 [sender.py:send():378] send: exit +2024-05-30 07:05:24,837 INFO SenderThread:900 [sender.py:send_exit():585] handling exit code: 1 +2024-05-30 07:05:24,837 INFO SenderThread:900 [sender.py:send_exit():587] handling runtime: 39 +2024-05-30 07:05:24,838 INFO SenderThread:900 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end +2024-05-30 07:05:24,838 INFO SenderThread:900 [sender.py:send_exit():593] send defer +2024-05-30 07:05:24,838 DEBUG HandlerThread:900 [handler.py:handle_request():158] handle_request: defer +2024-05-30 07:05:24,838 INFO HandlerThread:900 [handler.py:handle_request_defer():184] handle defer: 0 +2024-05-30 07:05:24,838 DEBUG SenderThread:900 [sender.py:send_request():405] send_request: defer +2024-05-30 07:05:24,838 INFO SenderThread:900 [sender.py:send_request_defer():609] handle sender defer: 0 +2024-05-30 07:05:24,838 INFO SenderThread:900 [sender.py:transition_state():613] send defer: 1 +2024-05-30 07:05:24,839 DEBUG HandlerThread:900 [handler.py:handle_request():158] handle_request: defer +2024-05-30 07:05:24,839 INFO HandlerThread:900 [handler.py:handle_request_defer():184] handle defer: 1 +2024-05-30 07:05:24,839 DEBUG SenderThread:900 [sender.py:send_request():405] send_request: defer +2024-05-30 07:05:24,839 INFO SenderThread:900 [sender.py:send_request_defer():609] handle sender defer: 1 +2024-05-30 07:05:24,839 INFO SenderThread:900 [sender.py:transition_state():613] send defer: 2 +2024-05-30 07:05:24,839 DEBUG HandlerThread:900 [handler.py:handle_request():158] handle_request: defer +2024-05-30 07:05:24,839 INFO HandlerThread:900 [handler.py:handle_request_defer():184] handle defer: 2 +2024-05-30 07:05:24,839 INFO HandlerThread:900 [system_monitor.py:finish():203] Stopping system monitor +2024-05-30 07:05:24,839 DEBUG SystemMonitor:900 [system_monitor.py:_start():172] Starting system metrics aggregation loop +2024-05-30 07:05:24,839 DEBUG SystemMonitor:900 [system_monitor.py:_start():179] Finished system metrics aggregation loop +2024-05-30 07:05:24,839 DEBUG SystemMonitor:900 [system_monitor.py:_start():183] Publishing last batch of metrics +2024-05-30 07:05:24,842 INFO HandlerThread:900 [interfaces.py:finish():200] Joined cpu monitor +2024-05-30 07:05:24,842 INFO HandlerThread:900 [interfaces.py:finish():200] Joined disk monitor +2024-05-30 07:05:24,842 INFO HandlerThread:900 [interfaces.py:finish():200] Joined memory monitor +2024-05-30 07:05:24,842 INFO HandlerThread:900 [interfaces.py:finish():200] Joined network monitor +2024-05-30 07:05:24,843 DEBUG SenderThread:900 [sender.py:send_request():405] send_request: defer +2024-05-30 07:05:24,843 INFO SenderThread:900 [sender.py:send_request_defer():609] handle sender defer: 2 +2024-05-30 07:05:24,843 INFO SenderThread:900 [sender.py:transition_state():613] send defer: 3 +2024-05-30 07:05:24,843 DEBUG SenderThread:900 [sender.py:send():378] send: stats +2024-05-30 07:05:24,844 DEBUG HandlerThread:900 [handler.py:handle_request():158] handle_request: defer +2024-05-30 07:05:24,844 INFO HandlerThread:900 [handler.py:handle_request_defer():184] handle defer: 3 +2024-05-30 07:05:24,844 DEBUG SenderThread:900 [sender.py:send_request():405] send_request: defer +2024-05-30 07:05:24,844 INFO SenderThread:900 [sender.py:send_request_defer():609] handle sender defer: 3 +2024-05-30 07:05:24,844 INFO SenderThread:900 [sender.py:transition_state():613] send defer: 4 +2024-05-30 07:05:24,844 DEBUG HandlerThread:900 [handler.py:handle_request():158] handle_request: defer +2024-05-30 07:05:24,844 INFO HandlerThread:900 [handler.py:handle_request_defer():184] handle defer: 4 +2024-05-30 07:05:24,844 DEBUG SenderThread:900 [sender.py:send_request():405] send_request: defer +2024-05-30 07:05:24,844 INFO SenderThread:900 [sender.py:send_request_defer():609] handle sender defer: 4 +2024-05-30 07:05:24,844 INFO SenderThread:900 [sender.py:transition_state():613] send defer: 5 +2024-05-30 07:05:24,845 DEBUG HandlerThread:900 [handler.py:handle_request():158] handle_request: defer +2024-05-30 07:05:24,845 INFO HandlerThread:900 [handler.py:handle_request_defer():184] handle defer: 5 +2024-05-30 07:05:24,845 DEBUG SenderThread:900 [sender.py:send():378] send: summary +2024-05-30 07:05:24,845 INFO SenderThread:900 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end +2024-05-30 07:05:24,846 DEBUG SenderThread:900 [sender.py:send_request():405] send_request: defer +2024-05-30 07:05:24,846 INFO SenderThread:900 [sender.py:send_request_defer():609] handle sender defer: 5 +2024-05-30 07:05:24,846 INFO SenderThread:900 [sender.py:transition_state():613] send defer: 6 +2024-05-30 07:05:24,846 DEBUG HandlerThread:900 [handler.py:handle_request():158] handle_request: defer +2024-05-30 07:05:24,846 INFO HandlerThread:900 [handler.py:handle_request_defer():184] handle defer: 6 +2024-05-30 07:05:24,846 DEBUG SenderThread:900 [sender.py:send_request():405] send_request: defer +2024-05-30 07:05:24,846 INFO SenderThread:900 [sender.py:send_request_defer():609] handle sender defer: 6 +2024-05-30 07:05:24,846 INFO SenderThread:900 [sender.py:transition_state():613] send defer: 7 +2024-05-30 07:05:24,846 DEBUG HandlerThread:900 [handler.py:handle_request():158] handle_request: status_report +2024-05-30 07:05:24,846 DEBUG HandlerThread:900 [handler.py:handle_request():158] handle_request: defer +2024-05-30 07:05:24,846 INFO HandlerThread:900 [handler.py:handle_request_defer():184] handle defer: 7 +2024-05-30 07:05:24,846 DEBUG SenderThread:900 [sender.py:send_request():405] send_request: defer +2024-05-30 07:05:24,846 INFO SenderThread:900 [sender.py:send_request_defer():609] handle sender defer: 7 +2024-05-30 07:05:25,819 INFO Thread-12 :900 [dir_watcher.py:_on_file_created():271] file/dir created: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240530_070444-c0riftte/files/wandb-summary.json +2024-05-30 07:05:25,837 DEBUG HandlerThread:900 [handler.py:handle_request():158] handle_request: poll_exit +2024-05-30 07:05:25,859 INFO SenderThread:900 [sender.py:transition_state():613] send defer: 8 +2024-05-30 07:05:25,859 DEBUG SenderThread:900 [sender.py:send_request():405] send_request: poll_exit +2024-05-30 07:05:25,859 DEBUG HandlerThread:900 [handler.py:handle_request():158] handle_request: defer +2024-05-30 07:05:25,859 INFO HandlerThread:900 [handler.py:handle_request_defer():184] handle defer: 8 +2024-05-30 07:05:25,859 DEBUG SenderThread:900 [sender.py:send_request():405] send_request: defer +2024-05-30 07:05:25,859 INFO SenderThread:900 [sender.py:send_request_defer():609] handle sender defer: 8 +2024-05-30 07:05:25,859 INFO SenderThread:900 [job_builder.py:build():432] Attempting to build job artifact +2024-05-30 07:05:25,860 INFO SenderThread:900 [job_builder.py:_get_source_type():576] no source found +2024-05-30 07:05:25,860 INFO SenderThread:900 [sender.py:transition_state():613] send defer: 9 +2024-05-30 07:05:25,860 DEBUG HandlerThread:900 [handler.py:handle_request():158] handle_request: defer +2024-05-30 07:05:25,860 INFO HandlerThread:900 [handler.py:handle_request_defer():184] handle defer: 9 +2024-05-30 07:05:25,860 DEBUG SenderThread:900 [sender.py:send_request():405] send_request: defer +2024-05-30 07:05:25,860 INFO SenderThread:900 [sender.py:send_request_defer():609] handle sender defer: 9 +2024-05-30 07:05:25,860 INFO SenderThread:900 [dir_watcher.py:finish():358] shutting down directory watcher +2024-05-30 07:05:26,821 INFO SenderThread:900 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240530_070444-c0riftte/files/output.log +2024-05-30 07:05:26,821 INFO SenderThread:900 [dir_watcher.py:finish():388] scan: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240530_070444-c0riftte/files +2024-05-30 07:05:26,821 INFO SenderThread:900 [dir_watcher.py:finish():402] scan save: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240530_070444-c0riftte/files/wandb-summary.json wandb-summary.json +2024-05-30 07:05:26,821 INFO SenderThread:900 [dir_watcher.py:finish():402] scan save: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240530_070444-c0riftte/files/config.yaml config.yaml +2024-05-30 07:05:26,824 INFO SenderThread:900 [dir_watcher.py:finish():402] scan save: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240530_070444-c0riftte/files/output.log output.log +2024-05-30 07:05:26,824 INFO SenderThread:900 [dir_watcher.py:finish():402] scan save: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240530_070444-c0riftte/files/wandb-metadata.json wandb-metadata.json +2024-05-30 07:05:26,824 INFO SenderThread:900 [dir_watcher.py:finish():402] scan save: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240530_070444-c0riftte/files/requirements.txt requirements.txt +2024-05-30 07:05:26,824 INFO SenderThread:900 [sender.py:transition_state():613] send defer: 10 +2024-05-30 07:05:26,825 DEBUG HandlerThread:900 [handler.py:handle_request():158] handle_request: defer +2024-05-30 07:05:26,825 INFO HandlerThread:900 [handler.py:handle_request_defer():184] handle defer: 10 +2024-05-30 07:05:26,825 DEBUG SenderThread:900 [sender.py:send_request():405] send_request: defer +2024-05-30 07:05:26,825 INFO SenderThread:900 [sender.py:send_request_defer():609] handle sender defer: 10 +2024-05-30 07:05:26,825 INFO SenderThread:900 [file_pusher.py:finish():169] shutting down file pusher +2024-05-30 07:05:26,837 DEBUG HandlerThread:900 [handler.py:handle_request():158] handle_request: poll_exit +2024-05-30 07:05:26,837 DEBUG SenderThread:900 [sender.py:send_request():405] send_request: poll_exit +2024-05-30 07:05:27,079 INFO wandb-upload_0:900 [upload_job.py:push():130] Uploaded file /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240530_070444-c0riftte/files/wandb-summary.json +2024-05-30 07:05:27,410 INFO wandb-upload_1:900 [upload_job.py:push():130] Uploaded file /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240530_070444-c0riftte/files/config.yaml +2024-05-30 07:05:27,456 INFO wandb-upload_3:900 [upload_job.py:push():130] Uploaded file /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240530_070444-c0riftte/files/requirements.txt +2024-05-30 07:05:27,485 INFO wandb-upload_2:900 [upload_job.py:push():130] Uploaded file /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240530_070444-c0riftte/files/output.log +2024-05-30 07:05:27,686 INFO Thread-11 (_thread_body):900 [sender.py:transition_state():613] send defer: 11 +2024-05-30 07:05:27,686 DEBUG HandlerThread:900 [handler.py:handle_request():158] handle_request: defer +2024-05-30 07:05:27,686 INFO HandlerThread:900 [handler.py:handle_request_defer():184] handle defer: 11 +2024-05-30 07:05:27,686 DEBUG SenderThread:900 [sender.py:send_request():405] send_request: defer +2024-05-30 07:05:27,686 INFO SenderThread:900 [sender.py:send_request_defer():609] handle sender defer: 11 +2024-05-30 07:05:27,686 INFO SenderThread:900 [file_pusher.py:join():175] waiting for file pusher +2024-05-30 07:05:27,686 INFO SenderThread:900 [sender.py:transition_state():613] send defer: 12 +2024-05-30 07:05:27,686 DEBUG HandlerThread:900 [handler.py:handle_request():158] handle_request: defer +2024-05-30 07:05:27,687 INFO HandlerThread:900 [handler.py:handle_request_defer():184] handle defer: 12 +2024-05-30 07:05:27,687 DEBUG SenderThread:900 [sender.py:send_request():405] send_request: defer +2024-05-30 07:05:27,687 INFO SenderThread:900 [sender.py:send_request_defer():609] handle sender defer: 12 +2024-05-30 07:05:27,687 INFO SenderThread:900 [file_stream.py:finish():601] file stream finish called +2024-05-30 07:05:27,773 INFO SenderThread:900 [file_stream.py:finish():605] file stream finish is done +2024-05-30 07:05:27,773 INFO SenderThread:900 [sender.py:transition_state():613] send defer: 13 +2024-05-30 07:05:27,773 DEBUG HandlerThread:900 [handler.py:handle_request():158] handle_request: defer +2024-05-30 07:05:27,773 INFO HandlerThread:900 [handler.py:handle_request_defer():184] handle defer: 13 +2024-05-30 07:05:27,774 DEBUG SenderThread:900 [sender.py:send_request():405] send_request: defer +2024-05-30 07:05:27,774 INFO SenderThread:900 [sender.py:send_request_defer():609] handle sender defer: 13 +2024-05-30 07:05:27,774 INFO SenderThread:900 [sender.py:transition_state():613] send defer: 14 +2024-05-30 07:05:27,774 DEBUG HandlerThread:900 [handler.py:handle_request():158] handle_request: defer +2024-05-30 07:05:27,774 INFO HandlerThread:900 [handler.py:handle_request_defer():184] handle defer: 14 +2024-05-30 07:05:27,774 DEBUG SenderThread:900 [sender.py:send():378] send: final +2024-05-30 07:05:27,774 DEBUG SenderThread:900 [sender.py:send():378] send: footer +2024-05-30 07:05:27,774 DEBUG SenderThread:900 [sender.py:send_request():405] send_request: defer +2024-05-30 07:05:27,774 INFO SenderThread:900 [sender.py:send_request_defer():609] handle sender defer: 14 +2024-05-30 07:05:27,775 DEBUG HandlerThread:900 [handler.py:handle_request():158] handle_request: poll_exit +2024-05-30 07:05:27,775 DEBUG SenderThread:900 [sender.py:send_request():405] send_request: poll_exit +2024-05-30 07:05:27,775 DEBUG HandlerThread:900 [handler.py:handle_request():158] handle_request: poll_exit +2024-05-30 07:05:27,775 DEBUG HandlerThread:900 [handler.py:handle_request():158] handle_request: server_info +2024-05-30 07:05:27,775 DEBUG HandlerThread:900 [handler.py:handle_request():158] handle_request: get_summary +2024-05-30 07:05:27,775 DEBUG HandlerThread:900 [handler.py:handle_request():158] handle_request: sampled_history +2024-05-30 07:05:27,775 DEBUG HandlerThread:900 [handler.py:handle_request():158] handle_request: internal_messages +2024-05-30 07:05:27,776 DEBUG SenderThread:900 [sender.py:send_request():405] send_request: poll_exit +2024-05-30 07:05:27,776 DEBUG SenderThread:900 [sender.py:send_request():405] send_request: server_info +2024-05-30 07:05:27,839 INFO MainThread:900 [wandb_run.py:_footer_history_summary_info():3994] rendering history +2024-05-30 07:05:27,839 INFO MainThread:900 [wandb_run.py:_footer_history_summary_info():4026] rendering summary +2024-05-30 07:05:27,839 INFO MainThread:900 [wandb_run.py:_footer_sync_info():3953] logging synced files +2024-05-30 07:05:27,840 DEBUG HandlerThread:900 [handler.py:handle_request():158] handle_request: shutdown +2024-05-30 07:05:27,840 INFO HandlerThread:900 [handler.py:finish():882] shutting down handler +2024-05-30 07:05:28,776 INFO WriterThread:900 [datastore.py:close():296] close: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240530_070444-c0riftte/run-c0riftte.wandb +2024-05-30 07:05:28,839 INFO SenderThread:900 [sender.py:finish():1545] shutting down sender +2024-05-30 07:05:28,839 INFO SenderThread:900 [file_pusher.py:finish():169] shutting down file pusher +2024-05-30 07:05:28,839 INFO SenderThread:900 [file_pusher.py:join():175] waiting for file pusher diff --git a/lm-evaluation-harness/wandb/run-20240530_070444-c0riftte/logs/debug.log b/lm-evaluation-harness/wandb/run-20240530_070444-c0riftte/logs/debug.log new file mode 100644 index 0000000000000000000000000000000000000000..b43036c8519c0afc796b666f956c1657ddd6f1b0 --- /dev/null +++ b/lm-evaluation-harness/wandb/run-20240530_070444-c0riftte/logs/debug.log @@ -0,0 +1,29 @@ +2024-05-30 07:04:44,757 INFO MainThread:744 [wandb_setup.py:_flush():76] Current SDK version is 0.17.0 +2024-05-30 07:04:44,757 INFO MainThread:744 [wandb_setup.py:_flush():76] Configure stats pid to 744 +2024-05-30 07:04:44,757 INFO MainThread:744 [wandb_setup.py:_flush():76] Loading settings from /root/.config/wandb/settings +2024-05-30 07:04:44,757 INFO MainThread:744 [wandb_setup.py:_flush():76] Loading settings from /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/settings +2024-05-30 07:04:44,757 INFO MainThread:744 [wandb_setup.py:_flush():76] Loading settings from environment variables: {} +2024-05-30 07:04:44,757 INFO MainThread:744 [wandb_setup.py:_flush():76] Applying setup settings: {'_disable_service': False} +2024-05-30 07:04:44,757 WARNING MainThread:744 [wandb_setup.py:_flush():76] Could not find program at -m lm_eval.__main__ +2024-05-30 07:04:44,757 INFO MainThread:744 [wandb_setup.py:_flush():76] Inferring run settings from compute environment: {'program_relpath': None, 'program': '-m lm_eval.__main__'} +2024-05-30 07:04:44,757 INFO MainThread:744 [wandb_setup.py:_flush():76] Applying login settings: {} +2024-05-30 07:04:44,757 INFO MainThread:744 [wandb_init.py:_log_setup():520] Logging user logs to /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240530_070444-c0riftte/logs/debug.log +2024-05-30 07:04:44,757 INFO MainThread:744 [wandb_init.py:_log_setup():521] Logging internal logs to /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240530_070444-c0riftte/logs/debug-internal.log +2024-05-30 07:04:44,757 INFO MainThread:744 [wandb_init.py:init():560] calling init triggers +2024-05-30 07:04:44,757 INFO MainThread:744 [wandb_init.py:init():567] wandb.init called with sweep_config: {} +config: {} +2024-05-30 07:04:44,757 INFO MainThread:744 [wandb_init.py:init():610] starting backend +2024-05-30 07:04:44,757 INFO MainThread:744 [wandb_init.py:init():614] setting up manager +2024-05-30 07:04:44,761 INFO MainThread:744 [backend.py:_multiprocessing_setup():105] multiprocessing start_methods=fork,spawn,forkserver, using: spawn +2024-05-30 07:04:44,762 INFO MainThread:744 [wandb_init.py:init():622] backend started and connected +2024-05-30 07:04:44,766 INFO MainThread:744 [wandb_init.py:init():711] updated telemetry +2024-05-30 07:04:44,774 INFO MainThread:744 [wandb_init.py:init():744] communicating run to backend with 90.0 second timeout +2024-05-30 07:04:45,272 INFO MainThread:744 [wandb_run.py:_on_init():2396] communicating current version +2024-05-30 07:04:45,426 INFO MainThread:744 [wandb_run.py:_on_init():2405] got version response +2024-05-30 07:04:45,426 INFO MainThread:744 [wandb_init.py:init():795] starting run threads in backend +2024-05-30 07:04:45,734 INFO MainThread:744 [wandb_run.py:_console_start():2374] atexit reg +2024-05-30 07:04:45,734 INFO MainThread:744 [wandb_run.py:_redirect():2229] redirect: wrap_raw +2024-05-30 07:04:45,734 INFO MainThread:744 [wandb_run.py:_redirect():2294] Wrapping output streams. +2024-05-30 07:04:45,734 INFO MainThread:744 [wandb_run.py:_redirect():2319] Redirects installed. +2024-05-30 07:04:45,750 INFO MainThread:744 [wandb_init.py:init():838] run started, returning control to user process +2024-05-30 07:05:28,840 WARNING MsgRouterThr:744 [router.py:message_loop():77] message_loop has been closed diff --git a/lm-evaluation-harness/wandb/run-20240530_070444-c0riftte/run-c0riftte.wandb b/lm-evaluation-harness/wandb/run-20240530_070444-c0riftte/run-c0riftte.wandb new file mode 100644 index 0000000000000000000000000000000000000000..4a98ac434e1f6cd0b895db1c5bf2dcc4848376be Binary files /dev/null and b/lm-evaluation-harness/wandb/run-20240530_070444-c0riftte/run-c0riftte.wandb differ diff --git a/lm-evaluation-harness/wandb/run-20240605_075207-j37ty43h/run-j37ty43h.wandb b/lm-evaluation-harness/wandb/run-20240605_075207-j37ty43h/run-j37ty43h.wandb new file mode 100644 index 0000000000000000000000000000000000000000..22a4d3b19fbeae7f6ff073b0df76923e3ec192e3 --- /dev/null +++ b/lm-evaluation-harness/wandb/run-20240605_075207-j37ty43h/run-j37ty43h.wandb @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:9beb933831b7a54c66c3868a6f42a7425b8de80088164e855a22dae7c370b8f4 +size 1369600 diff --git a/lm-evaluation-harness/wandb/run-20240608_111026-9apxn9eo/run-9apxn9eo.wandb b/lm-evaluation-harness/wandb/run-20240608_111026-9apxn9eo/run-9apxn9eo.wandb new file mode 100644 index 0000000000000000000000000000000000000000..ee79b3bb9846f8cfba725b81e906f950140903a3 --- /dev/null +++ b/lm-evaluation-harness/wandb/run-20240608_111026-9apxn9eo/run-9apxn9eo.wandb @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:65ecec43c786b7efc4ab9e84b03811946843ec0663763820a708248893e283e7 +size 1333167 diff --git a/lm-evaluation-harness/wandb/run-20240608_134753-jc54oi3l/run-jc54oi3l.wandb b/lm-evaluation-harness/wandb/run-20240608_134753-jc54oi3l/run-jc54oi3l.wandb new file mode 100644 index 0000000000000000000000000000000000000000..1bc3ee900c32940289624ee57f5ce25802126f60 --- /dev/null +++ b/lm-evaluation-harness/wandb/run-20240608_134753-jc54oi3l/run-jc54oi3l.wandb @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:145e6afe560cb12f8bb3bfbf40292aae4939e5524d23ed3458d9dc97bed3e839 +size 1294651 diff --git a/lm-evaluation-harness/wandb/run-20240608_190333-82mnef5m/run-82mnef5m.wandb b/lm-evaluation-harness/wandb/run-20240608_190333-82mnef5m/run-82mnef5m.wandb new file mode 100644 index 0000000000000000000000000000000000000000..6eb543220d7f7890364fe3dedd479944411ece76 --- /dev/null +++ b/lm-evaluation-harness/wandb/run-20240608_190333-82mnef5m/run-82mnef5m.wandb @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:b7ebe86f007e5050375f011ab357e13e47d3b4bc644db416495a6bc00ed6f100 +size 1292144