diff --git a/lm-evaluation-harness/tests/models/test_gguf.py b/lm-evaluation-harness/tests/models/test_gguf.py new file mode 100644 index 0000000000000000000000000000000000000000..186b2305e6afa2a215961f1aa98c2f8b87cdd4e2 --- /dev/null +++ b/lm-evaluation-harness/tests/models/test_gguf.py @@ -0,0 +1,152 @@ +import hashlib +import json +import os +import pickle +import unittest +from unittest.mock import patch + +from lm_eval.api.instance import Instance +from lm_eval.models.gguf import GGUFLM + + +base_url = "https://matthoffner-ggml-llm-api.hf.space" + + +def gguf_completion_mock(base_url=None, **kwargs): + # Generate a hash from the parameters + hash_kwargs = {"base_url": base_url, **kwargs} + hash = hashlib.sha256( + json.dumps(hash_kwargs, sort_keys=True).encode("utf-8") + ).hexdigest() + + fname = f"./tests/testdata/gguf_test_{hash}.pkl" + + if os.path.exists(fname): + with open(fname, "rb") as fh: + return pickle.load(fh) + else: + print("The file does not exist, attempting to write...") + if "stop" in kwargs: + result = { + "choices": [ + { + "text": f"generated text until {kwargs['stop']}", + "logprobs": {"token_logprobs": [-1.2345], "text_offset": 0}, + "finish_reason": "length", + } + ] + } + else: + # generated with # curl -X 'POST' 'http://localhost:8000/v1/completions' -H 'accept: application/json' -H 'Content-Type: application/json' -d '{"prompt": "string", "logprobs": 10, "temperature": 0.0, "max_tokens": 1, "echo": true}' + result = { + "id": "cmpl-4023976b-bc6a-43b0-a5a9-629f4216c7f3", + "object": "text_completion", + "created": 1700511361, + "model": "../llama-2-7b.Q8_0.gguf", + "choices": [ + { + "text": "string(", + "index": 0, + "logprobs": { + "text_offset": [0, 7], + "token_logprobs": [None, -1.033263319857306], + "tokens": [" string", "("], + "top_logprobs": [ + None, + { + "(": -1.033263319857306, + "[]": -2.6530743779017394, + ".": -3.0377145947291324, + "\n": -3.0399156750513976, + "_": -3.510376089937872, + " =": -3.6957918347193663, + ",": -3.9309459866358702, + " of": -4.2834550083949035, + '("': -4.322762841112799, + "()": -4.426229113466925, + }, + ], + }, + "finish_reason": "length", + } + ], + "usage": { + "prompt_tokens": 2, + "completion_tokens": 1, + "total_tokens": 3, + }, + } + + try: + os.makedirs(os.path.dirname(fname), exist_ok=True) + print("Writing file at", fname) + with open(fname, "wb") as fh: + pickle.dump(result, fh) + print("File written successfully") + except Exception as e: + print("File writing failed:", e) + + return result + + +class GGUFLMTest(unittest.TestCase): + @patch( + "lm_eval.models.gguf.GGUFLM.gguf_completion", side_effect=gguf_completion_mock + ) + def test_loglikelihood(self, gguf_completion_mock): + lm = GGUFLM(base_url) + + # Test loglikelihood + requests = [ + Instance( + request_type="loglikelihood", + doc=args, + arguments=args, + idx=i, + ) + for i, args in enumerate([("str", "ing"), ("str", "ing")]) + ] + res = lm.loglikelihood(requests) + + # Assert the loglikelihood response is correct + expected_res = [(logprob, True) for logprob in [0, 0]] + self.assertEqual(res, expected_res) + + @patch( + "lm_eval.models.gguf.GGUFLM.gguf_completion", side_effect=gguf_completion_mock + ) + def test_generate_until(self, gguf_completion_mock): + lm = GGUFLM(base_url) + + # Test generate_until + requests = [ + Instance( + request_type="generate_until", + doc={"input": doc}, + arguments=(doc, {"until": stop}), + idx=i, + ) + for i, (doc, stop) in enumerate([("input1", "stop1"), ("input2", "stop2")]) + ] + + res = lm.generate_until(requests) + + # Assert the generate_until response is correct + expected_res = ["generated text until stop1", "generated text until stop2"] + self.assertEqual(res, expected_res) + + # @patch('lm_eval.models.gguf.GGUFLM.gguf_completion', side_effect=gguf_completion_mock) + # def test_loglikelihood_rolling(self, gguf_completion_mock): + # lm = GGUFLM(base_url) + + # # Test loglikelihood_rolling + # requests = ["input1", "input2"] + # res = lm.loglikelihood_rolling(requests) + + # # Assert the loglikelihood_rolling response is correct + # expected_res = [(-1.2345, True), (-1.2345, True)] + # self.assertEqual(res, expected_res) + + +if __name__ == "__main__": + unittest.main() diff --git a/lm-evaluation-harness/tests/models/test_huggingface.py b/lm-evaluation-harness/tests/models/test_huggingface.py new file mode 100644 index 0000000000000000000000000000000000000000..efef5a9b7ffa4267842c32f4d203e1d630a24d68 --- /dev/null +++ b/lm-evaluation-harness/tests/models/test_huggingface.py @@ -0,0 +1,143 @@ +from __future__ import annotations + +import sys +from pathlib import Path + +import numpy as np +import torch + +import lm_eval.tasks as tasks +from lm_eval.api.instance import Instance +from lm_eval.models.huggingface import HFLM + + +task_manager = tasks.TaskManager() + + +class Test_HFLM: + torch.use_deterministic_algorithms(True) + task_list = task_manager.load_task_or_group(["arc_easy", "gsm8k", "wikitext"]) + version_minor = sys.version_info.minor + multiple_choice_task = task_list["arc_easy"] # type: ignore + multiple_choice_task.build_all_requests(limit=10, rank=0, world_size=1) + MULTIPLE_CH: list[Instance] = multiple_choice_task.instances + generate_until_task = task_list["gsm8k"] # type: ignore + generate_until_task._config.generation_kwargs["max_gen_toks"] = 10 + generate_until_task.build_all_requests(limit=10, rank=0, world_size=1) + generate_until: list[Instance] = generate_until_task.instances + rolling_task = task_list["wikitext"] # type: ignore + rolling_task.build_all_requests(limit=10, rank=0, world_size=1) + ROLLING: list[Instance] = rolling_task.instances + + MULTIPLE_CH_RES = [ + -41.902435302734375, + -42.939308166503906, + -33.914180755615234, + -37.07139205932617, + -22.95258331298828, + -20.342208862304688, + -14.818366050720215, + -27.942853927612305, + -15.80704116821289, + -15.936427116394043, + -13.052018165588379, + -18.04828453063965, + -13.345029830932617, + -13.366025924682617, + -12.127134323120117, + -11.872495651245117, + -47.10598373413086, + -47.76410675048828, + -36.4406852722168, + -50.0289421081543, + -16.72093963623047, + -18.535587310791016, + -26.46993637084961, + -20.355995178222656, + -17.757919311523438, + -21.80595588684082, + -33.1990852355957, + -39.28636932373047, + -14.759679794311523, + -16.753942489624023, + -11.486852645874023, + -15.42177677154541, + -13.15798282623291, + -15.887393951416016, + -15.28614616394043, + -12.339089393615723, + -44.59441375732422, + -55.40888214111328, + -52.70050811767578, + -56.25089645385742, + ] + generate_until_RES = [ + " The average of $2.50 each is $", + " A robe takes 2 bolts of blue fiber and half", + " $50,000 in repairs.\n\nQuestion", + " He runs 1 sprint 3 times a week.", + " They feed each of her chickens three cups of mixed", + " The price of the glasses is $5, but", + " The total percentage of students who said they like to", + " Carla is downloading a 200 GB file. Normally", + " John drives for 3 hours at a speed of 60", + " Eliza sells 4 tickets to 5 friends so she", + ] + ROLLING_RES = [ + -3603.6328125, + -19779.23974609375, + -8834.16455078125, + -27967.591796875, + -7636.794982910156, + -9491.93505859375, + -41043.4248046875, + -8397.689819335938, + -45969.47155761719, + -7158.90625, + ] + LM = HFLM(pretrained="EleutherAI/pythia-70m", device="cpu", dtype="float32") + + def test_logliklihood(self) -> None: + res = self.LM.loglikelihood(self.MULTIPLE_CH) + _RES, _res = self.MULTIPLE_CH_RES, [r[0] for r in res] + # log samples to CI + dir_path = Path("test_logs") + dir_path.mkdir(parents=True, exist_ok=True) + + file_path = dir_path / f"outputs_log_{self.version_minor}.txt" + file_path = file_path.resolve() + with open(file_path, "w") as f: + f.write("\n".join(str(x) for x in _res)) + assert np.allclose(_res, _RES, atol=1e-2) + # check indices for Multiple Choice + argmax_RES, argmax_res = ( + np.argmax(np.array(_RES).reshape(-1, 4), axis=1), + np.argmax(np.array(_res).reshape(-1, 4), axis=1), + ) + assert (argmax_RES == argmax_res).all() + + def test_generate_until(self) -> None: + res = self.LM.generate_until(self.generate_until) + assert res == self.generate_until_RES + + def test_logliklihood_rolling(self) -> None: + res = self.LM.loglikelihood_rolling(self.ROLLING) + assert np.allclose(res, self.ROLLING_RES, atol=1e-1) + + def test_toc_encode(self) -> None: + res = self.LM.tok_encode("foo bar") + assert res == [12110, 2534] + + def test_toc_decode(self) -> None: + res = self.LM.tok_decode([12110, 2534]) + assert res == "foo bar" + + def test_batch_encode(self) -> None: + res = self.LM.tok_batch_encode(["foo bar", "bar foo"])[0].tolist() + assert res == [[12110, 2534], [2009, 17374]] + + def test_model_generate(self) -> None: + context = self.LM.tok_batch_encode(["foo bar"])[0] + res = self.LM._model_generate(context, max_length=10, stop=["\n\n"]) + res = self.LM.tok_decode(res[0]) + assert res == "foo bar\n!info bar" diff --git a/lm-evaluation-harness/tests/models/test_neuralmagic.py b/lm-evaluation-harness/tests/models/test_neuralmagic.py new file mode 100644 index 0000000000000000000000000000000000000000..c7446173ffe775e9f61cbe5065c2feb11f1dc4b2 --- /dev/null +++ b/lm-evaluation-harness/tests/models/test_neuralmagic.py @@ -0,0 +1,61 @@ +import pytest + +import lm_eval.evaluator as evaluator +from lm_eval.api.registry import get_model + + +SPARSEML_MODELS_TASKS = [ + # loglikelihood + ("facebook/opt-125m", "lambada_openai"), + # loglikelihood_rolling + ("hf-internal-testing/tiny-random-gpt2", "wikitext"), + # generate_until + ("mgoin/tiny-random-llama-2-quant", "gsm8k"), +] + +DEEPSPARSE_MODELS_TASKS = [ + # loglikelihood + ("hf:mgoin/llama2.c-stories15M-quant-ds", "lambada_openai"), + # loglikelihood_rolling (not supported yet) + # ("hf:mgoin/llama2.c-stories15M-quant-ds", "wikitext"), + # generate_until + ("hf:mgoin/llama2.c-stories15M-quant-ds", "gsm8k"), +] + + +@pytest.mark.parametrize("model_id,task", SPARSEML_MODELS_TASKS) +def test_sparseml_eval(model_id, task): + lm = get_model("sparseml").create_from_arg_string( + f"pretrained={model_id}", + { + "batch_size": 1, + "device": "cpu", + "dtype": "float32", + }, + ) + + limit = 5 + evaluator.simple_evaluate( + model=lm, + tasks=[task], + num_fewshot=0, + limit=limit, + ) + + +@pytest.mark.parametrize("model_id,task", DEEPSPARSE_MODELS_TASKS) +def test_deepsparse_eval(model_id, task): + lm = get_model("deepsparse").create_from_arg_string( + f"pretrained={model_id}", + { + "batch_size": 1, + }, + ) + + limit = 5 + evaluator.simple_evaluate( + model=lm, + tasks=[task], + num_fewshot=0, + limit=limit, + ) diff --git a/lm-evaluation-harness/tests/models/test_neuron_optimum.py b/lm-evaluation-harness/tests/models/test_neuron_optimum.py new file mode 100644 index 0000000000000000000000000000000000000000..564d52303968e210439a7931f012487a959a367f --- /dev/null +++ b/lm-evaluation-harness/tests/models/test_neuron_optimum.py @@ -0,0 +1,26 @@ +import pytest +import torch + +from lm_eval.models.neuron_optimum import wrap_constant_batch_size + + +def test_wrap_constant_batch_size(): + class Tester: + def __init__(self, batch_size): + self.batch_size = batch_size + + @wrap_constant_batch_size + def test_constant_batch_size(self, inputs): + assert len(inputs) == self.batch_size + return inputs + + batch_size_test = 8 + for i in range(1, batch_size_test + 1): + tensor = torch.ones([i, 2, 2]) + out = Tester(batch_size=batch_size_test).test_constant_batch_size(tensor) + torch.testing.assert_allclose(out, tensor) + + with pytest.raises(ValueError): + Tester(batch_size=batch_size_test).test_constant_batch_size( + torch.ones([batch_size_test + 1, 2, 2]) + ) diff --git a/lm-evaluation-harness/tests/models/test_openvino.py b/lm-evaluation-harness/tests/models/test_openvino.py new file mode 100644 index 0000000000000000000000000000000000000000..34fc416a6ead50fb25f6ab659aa77b23ec58e068 --- /dev/null +++ b/lm-evaluation-harness/tests/models/test_openvino.py @@ -0,0 +1,73 @@ +import random +import tempfile + +import pytest +from optimum.intel import OVModelForCausalLM +from transformers import AutoTokenizer + +import lm_eval.evaluator as evaluator +from lm_eval.api.registry import get_model + + +SUPPORTED_ARCHITECTURES_TASKS = { + "facebook/opt-125m": "lambada_openai", + "hf-internal-testing/tiny-random-gpt2": "wikitext", +} + + +@pytest.mark.parametrize("model_id,task", SUPPORTED_ARCHITECTURES_TASKS.items()) +def test_evaluator(model_id, task): + with tempfile.TemporaryDirectory() as tmpdirname: + model = OVModelForCausalLM.from_pretrained( + model_id, export=True, use_cache=True + ) + model.save_pretrained(tmpdirname) + tokenizer = AutoTokenizer.from_pretrained(model_id) + tokenizer.save_pretrained(tmpdirname) + + lm = get_model("openvino").create_from_arg_string( + f"pretrained={tmpdirname}", + { + "batch_size": 1, + "device": "cpu", + }, + ) + + def ll_fn(reqs): + for ctx, cont in [req.args for req in reqs]: + if len(ctx) == 0: + continue + # space convention + assert ctx[-1] != " " + assert cont[0] == " " or ctx[-1] == "\n" + + res = [] + + random.seed(42) + for _ in reqs: + res.append((-random.random(), False)) + + return res + + def ll_perp_fn(reqs): + for (string,) in [req.args for req in reqs]: + assert isinstance(string, str) + + res = [] + random.seed(42) + for _ in reqs: + res.append(-random.random()) + + return res + + lm.loglikelihood = ll_fn + lm.loglikelihood_rolling = ll_perp_fn + + limit = 10 + evaluator.simple_evaluate( + model=lm, + tasks=[task], + num_fewshot=0, + limit=limit, + bootstrap_iters=10, + ) diff --git a/lm-evaluation-harness/tests/models/test_vllm.py b/lm-evaluation-harness/tests/models/test_vllm.py new file mode 100644 index 0000000000000000000000000000000000000000..280c24c119900fcf5c3c41d517207bf8720795ef --- /dev/null +++ b/lm-evaluation-harness/tests/models/test_vllm.py @@ -0,0 +1,51 @@ +from typing import List + +import pytest +import torch + +import lm_eval.tasks as tasks +from lm_eval.api.instance import Instance + + +task_manager = tasks.TaskManager() + + +@pytest.mark.skip(reason="requires CUDA") +class TEST_VLLM: + vllm = pytest.importorskip("vllm") + try: + from lm_eval.models.vllm_causallms import VLLM + + LM = VLLM(pretrained="EleutherAI/pythia-70m") + except ModuleNotFoundError: + pass + torch.use_deterministic_algorithms(True) + task_list = task_manager.load_task_or_group(["arc_easy", "gsm8k", "wikitext"]) + multiple_choice_task = task_list["arc_easy"] # type: ignore + multiple_choice_task.build_all_requests(limit=10, rank=0, world_size=1) + MULTIPLE_CH: List[Instance] = multiple_choice_task.instances + generate_until_task = task_list["gsm8k"] # type: ignore + generate_until_task._config.generation_kwargs["max_gen_toks"] = 10 + generate_until_task.build_all_requests(limit=10, rank=0, world_size=1) + generate_until: List[Instance] = generate_until_task.instances + rolling_task = task_list["wikitext"] # type: ignore + rolling_task.build_all_requests(limit=10, rank=0, world_size=1) + ROLLING: List[Instance] = rolling_task.instances + + # TODO: make proper tests + def test_logliklihood(self) -> None: + res = self.LM.loglikelihood(self.MULTIPLE_CH) + assert len(res) == len(self.MULTIPLE_CH) + for x in res: + assert isinstance(x[0], float) + + def test_generate_until(self) -> None: + res = self.LM.generate_until(self.generate_until) + assert len(res) == len(self.generate_until) + for x in res: + assert isinstance(x, str) + + def test_logliklihood_rolling(self) -> None: + res = self.LM.loglikelihood_rolling(self.ROLLING) + for x in res: + assert isinstance(x, float) diff --git a/lm-evaluation-harness/tests/testdata/blimp_distractor_agreement_relational_noun-v0-loglikelihood b/lm-evaluation-harness/tests/testdata/blimp_distractor_agreement_relational_noun-v0-loglikelihood new file mode 100644 index 0000000000000000000000000000000000000000..f926cf3d4b7c1fb9fe3662b21754329ecc15ee2f --- /dev/null +++ b/lm-evaluation-harness/tests/testdata/blimp_distractor_agreement_relational_noun-v0-loglikelihood @@ -0,0 +1 @@ +8aab641bd5933f84f46a14f5c1208a3c855cace7e67b44abcd5aff8fec96717d \ No newline at end of file diff --git a/lm-evaluation-harness/tests/testdata/blimp_existential_there_subject_raising-v0-loglikelihood b/lm-evaluation-harness/tests/testdata/blimp_existential_there_subject_raising-v0-loglikelihood new file mode 100644 index 0000000000000000000000000000000000000000..925e5b4680b003be07aad25d99c377b16c5c18e0 --- /dev/null +++ b/lm-evaluation-harness/tests/testdata/blimp_existential_there_subject_raising-v0-loglikelihood @@ -0,0 +1 @@ +9b324b28ae3e1b5d49ecf4b7b2a16c7bbc8ff38d000cf216fab75df633da2084 \ No newline at end of file diff --git a/lm-evaluation-harness/tests/testdata/blimp_expletive_it_object_raising-v0-res.json b/lm-evaluation-harness/tests/testdata/blimp_expletive_it_object_raising-v0-res.json new file mode 100644 index 0000000000000000000000000000000000000000..735dc09826d056ed20a40b8bd9ccf54b434d05a8 --- /dev/null +++ b/lm-evaluation-harness/tests/testdata/blimp_expletive_it_object_raising-v0-res.json @@ -0,0 +1 @@ +{"results": {"blimp_expletive_it_object_raising": {"acc": 0.485, "acc_stderr": 0.0158121796418149}}, "versions": {"blimp_expletive_it_object_raising": 0}} \ No newline at end of file diff --git a/lm-evaluation-harness/tests/testdata/crows_pairs_english_race_color-v0-res.json b/lm-evaluation-harness/tests/testdata/crows_pairs_english_race_color-v0-res.json new file mode 100644 index 0000000000000000000000000000000000000000..75d356522edf08f93f03d3ba37ed323d39f5b35e --- /dev/null +++ b/lm-evaluation-harness/tests/testdata/crows_pairs_english_race_color-v0-res.json @@ -0,0 +1 @@ +{"results": {"crows_pairs_english_race_color": {"likelihood_difference": 0.3322827903840805, "likelihood_difference_stderr": 0.01019838186372816, "pct_stereotype": 0.4822834645669291, "pct_stereotype_stderr": 0.022191835500120254}}, "versions": {"crows_pairs_english_race_color": 0}} \ No newline at end of file diff --git a/lm-evaluation-harness/tests/testdata/hellaswag-v0-res.json b/lm-evaluation-harness/tests/testdata/hellaswag-v0-res.json new file mode 100644 index 0000000000000000000000000000000000000000..6be94a640950b2451775fddccbf80060c4a673b0 --- /dev/null +++ b/lm-evaluation-harness/tests/testdata/hellaswag-v0-res.json @@ -0,0 +1 @@ +{"results": {"hellaswag": {"acc": 0.24965146385182235, "acc_norm": 0.24756024696275641, "acc_norm_stderr": 0.004307128573285236, "acc_stderr": 0.004319267432460666}}, "versions": {"hellaswag": 0}} \ No newline at end of file diff --git a/lm-evaluation-harness/tests/testdata/hendrycksTest-econometrics-v0-res.json b/lm-evaluation-harness/tests/testdata/hendrycksTest-econometrics-v0-res.json new file mode 100644 index 0000000000000000000000000000000000000000..4656fac3c3026ec7e137ce8f49e4796fefe5e24f --- /dev/null +++ b/lm-evaluation-harness/tests/testdata/hendrycksTest-econometrics-v0-res.json @@ -0,0 +1 @@ +{"results": {"hendrycksTest-econometrics": {"acc": 0.24561403508771928, "acc_norm": 0.24561403508771928, "acc_norm_stderr": 0.04049339297748142, "acc_stderr": 0.040493392977481425}}, "versions": {"hendrycksTest-econometrics": 0}} \ No newline at end of file diff --git a/lm-evaluation-harness/tests/testdata/hendrycksTest-us_foreign_policy-v0-res.json b/lm-evaluation-harness/tests/testdata/hendrycksTest-us_foreign_policy-v0-res.json new file mode 100644 index 0000000000000000000000000000000000000000..1077380de88cb9ce23894ce31fbbeceea90f2079 --- /dev/null +++ b/lm-evaluation-harness/tests/testdata/hendrycksTest-us_foreign_policy-v0-res.json @@ -0,0 +1 @@ +{"results": {"hendrycksTest-us_foreign_policy": {"acc": 0.2, "acc_norm": 0.24, "acc_norm_stderr": 0.04292346959909283, "acc_stderr": 0.040201512610368445}}, "versions": {"hendrycksTest-us_foreign_policy": 0}} \ No newline at end of file diff --git a/lm-evaluation-harness/tests/testdata/multirc-v1-res.json b/lm-evaluation-harness/tests/testdata/multirc-v1-res.json new file mode 100644 index 0000000000000000000000000000000000000000..938141bbb888f55c3aa2786868c28925ac3fd123 --- /dev/null +++ b/lm-evaluation-harness/tests/testdata/multirc-v1-res.json @@ -0,0 +1 @@ +{"results": {"multirc": {"acc": 0.046169989506820566, "acc_stderr": 0.006801377886208738}}, "versions": {"multirc": 1}} \ No newline at end of file diff --git a/lm-evaluation-harness/tests/testdata/pile_arxiv-v0-loglikelihood_rolling b/lm-evaluation-harness/tests/testdata/pile_arxiv-v0-loglikelihood_rolling new file mode 100644 index 0000000000000000000000000000000000000000..3aa1d8c7349449271fbd81fbbc06fde47a116028 --- /dev/null +++ b/lm-evaluation-harness/tests/testdata/pile_arxiv-v0-loglikelihood_rolling @@ -0,0 +1 @@ +814f9954e44368559602c00f7e85fa3971acdfd0315f508ec7df6318a79c55ec \ No newline at end of file diff --git a/lm-evaluation-harness/tests/testdata/pile_philpapers-v1-res.json b/lm-evaluation-harness/tests/testdata/pile_philpapers-v1-res.json new file mode 100644 index 0000000000000000000000000000000000000000..5a2f77678abc264edf433a4eb98da08fc20b1dfc --- /dev/null +++ b/lm-evaluation-harness/tests/testdata/pile_philpapers-v1-res.json @@ -0,0 +1 @@ +{"results": {"pile_philpapers": {"bits_per_byte": 9.004690592465457e-06, "byte_perplexity": 1.0000062415953748, "word_perplexity": 1.0000409888564146}}, "versions": {"pile_philpapers": 1}} \ No newline at end of file diff --git a/lm-evaluation-harness/tests/testdata/pile_pubmed-central-v0-loglikelihood_rolling b/lm-evaluation-harness/tests/testdata/pile_pubmed-central-v0-loglikelihood_rolling new file mode 100644 index 0000000000000000000000000000000000000000..283109f32e0aac45adcbc90c7c8fb41114e7771f --- /dev/null +++ b/lm-evaluation-harness/tests/testdata/pile_pubmed-central-v0-loglikelihood_rolling @@ -0,0 +1 @@ +40b39d120d99a145690444e86acc3e3e24d41e6e0538a75e26929ad84926e5e0 \ No newline at end of file diff --git a/lm-evaluation-harness/tests/testdata/pile_ubuntu-irc-v1-loglikelihood_rolling b/lm-evaluation-harness/tests/testdata/pile_ubuntu-irc-v1-loglikelihood_rolling new file mode 100644 index 0000000000000000000000000000000000000000..ce041998635643ee17aace3105b227ef0746917e --- /dev/null +++ b/lm-evaluation-harness/tests/testdata/pile_ubuntu-irc-v1-loglikelihood_rolling @@ -0,0 +1 @@ +4eb69e314f0864ec8890e2323d7e76f8a8309692c4f090e2b41bf4be681a811d \ No newline at end of file diff --git a/lm-evaluation-harness/wandb/debug-cli.root.log b/lm-evaluation-harness/wandb/debug-cli.root.log new file mode 100644 index 0000000000000000000000000000000000000000..e69de29bb2d1d6434b8b29ae775ad8c2e48c5391 diff --git a/lm-evaluation-harness/wandb/debug-internal.log b/lm-evaluation-harness/wandb/debug-internal.log new file mode 100644 index 0000000000000000000000000000000000000000..e7d94ea3c787b2ac4a96d0c3a6a65729617a95f3 --- /dev/null +++ b/lm-evaluation-harness/wandb/debug-internal.log @@ -0,0 +1,3651 @@ +2024-06-08 19:03:33,780 INFO StreamThr :30410 [internal.py:wandb_internal():85] W&B internal server running at pid: 30410, started at: 2024-06-08 19:03:33.777669 +2024-06-08 19:03:33,784 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: status +2024-06-08 19:03:33,785 INFO WriterThread:30410 [datastore.py:open_for_write():87] open: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_190333-82mnef5m/run-82mnef5m.wandb +2024-06-08 19:03:33,787 DEBUG SenderThread:30410 [sender.py:send():379] send: header +2024-06-08 19:03:33,791 DEBUG SenderThread:30410 [sender.py:send():379] send: run +2024-06-08 19:03:33,989 INFO SenderThread:30410 [dir_watcher.py:__init__():211] watching files in: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_190333-82mnef5m/files +2024-06-08 19:03:33,989 INFO SenderThread:30410 [sender.py:_start_run_threads():1188] run started: 82mnef5m with start time 1717873413.777759 +2024-06-08 19:03:33,992 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: check_version +2024-06-08 19:03:33,992 DEBUG SenderThread:30410 [sender.py:send_request():406] send_request: check_version +2024-06-08 19:03:34,058 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: run_start +2024-06-08 19:03:34,061 DEBUG HandlerThread:30410 [system_info.py:__init__():26] System info init +2024-06-08 19:03:34,061 DEBUG HandlerThread:30410 [system_info.py:__init__():41] System info init done +2024-06-08 19:03:34,061 INFO HandlerThread:30410 [system_monitor.py:start():194] Starting system monitor +2024-06-08 19:03:34,061 INFO SystemMonitor:30410 [system_monitor.py:_start():158] Starting system asset monitoring threads +2024-06-08 19:03:34,061 INFO HandlerThread:30410 [system_monitor.py:probe():214] Collecting system info +2024-06-08 19:03:34,068 INFO SystemMonitor:30410 [interfaces.py:start():188] Started cpu monitoring +2024-06-08 19:03:34,068 INFO SystemMonitor:30410 [interfaces.py:start():188] Started disk monitoring +2024-06-08 19:03:34,074 INFO SystemMonitor:30410 [interfaces.py:start():188] Started memory monitoring +2024-06-08 19:03:34,075 INFO SystemMonitor:30410 [interfaces.py:start():188] Started network monitoring +2024-06-08 19:03:34,154 DEBUG HandlerThread:30410 [system_info.py:probe():152] Probing system +2024-06-08 19:03:34,157 DEBUG HandlerThread:30410 [system_info.py:_probe_git():137] Probing git +2024-06-08 19:03:34,167 ERROR HandlerThread:30410 [gitlib.py:root():92] git root error: Cmd('git') failed due to: exit code(128) + cmdline: git rev-parse --show-toplevel + stderr: 'fatal: detected dubious ownership in repository at '/mnt/weka/peacock/idc/cronscript/lm-evaluation-harness' +To add an exception for this directory, call: + + git config --global --add safe.directory /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness' +2024-06-08 19:03:34,167 DEBUG HandlerThread:30410 [system_info.py:_probe_git():145] Probing git done +2024-06-08 19:03:34,167 DEBUG HandlerThread:30410 [system_info.py:probe():200] Probing system done +2024-06-08 19:03:34,167 DEBUG HandlerThread:30410 [system_monitor.py:probe():223] {'os': 'Linux-5.15.0-92-generic-x86_64-with-glibc2.35', 'python': '3.10.12', 'heartbeatAt': '2024-06-08T19:03:34.154505', 'startedAt': '2024-06-08T19:03:33.753786', 'docker': None, 'cuda': None, 'args': ('--model', 'hf', '--model_args', 'pretrained=/mnt/weka/peacock/experiments/llama/eval/checkpoint-enhibn-updated/llamav2-3b/hf/global_step240000,tokenizer=/mnt/weka/peacock/tokenization/trained-tokenizer/enhiben_50k_hf/ConvertedTokenizer', '--tasks', 'winogrande,sst2,mrpc,arc_easy,copa,piqa,boolq,indic_arc_easy_hi,indic_arc_challenge_hi,indic_boolq_hi', '--batch_size', 'auto', '--wandb_args', 'project=engl-hi-eval,name=globalstep_240000,group=exp2'), 'state': 'running', 'program': '-m lm_eval.__main__', 'codePathLocal': None, 'git': {'remote': 'https://github.com/EleutherAI/lm-evaluation-harness', 'commit': None}, 'email': None, 'root': '/mnt/weka/peacock/idc/cronscript/lm-evaluation-harness', 'host': 'peacock-evaluation-debug-worker-0', 'username': 'root', 'executable': '/usr/bin/python3', 'cpu_count': 80, 'cpu_count_logical': 160, 'cpu_freq': {'current': 2327.499975, 'min': 800.0, 'max': 3400.0}, 'cpu_freq_per_core': [{'current': 3311.804, 'min': 800.0, 'max': 3400.0}, {'current': 3311.816, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 3299.996, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 3311.792, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}], 'disk': {'/': {'total': 877.6341285705566, 'used': 214.68004989624023}}, 'memory': {'total': 1007.4379501342773}} +2024-06-08 19:03:34,168 INFO HandlerThread:30410 [system_monitor.py:probe():224] Finished collecting system info +2024-06-08 19:03:34,168 INFO HandlerThread:30410 [system_monitor.py:probe():227] Publishing system info +2024-06-08 19:03:34,171 INFO HandlerThread:30410 [system_monitor.py:probe():229] Finished publishing system info +2024-06-08 19:03:34,176 DEBUG SenderThread:30410 [sender.py:send():379] send: files +2024-06-08 19:03:34,176 INFO SenderThread:30410 [sender.py:_save_file():1454] saving file wandb-metadata.json with policy now +2024-06-08 19:03:34,349 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: python_packages +2024-06-08 19:03:34,349 DEBUG SenderThread:30410 [sender.py:send_request():406] send_request: python_packages +2024-06-08 19:03:34,351 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: stop_status +2024-06-08 19:03:34,352 DEBUG SenderThread:30410 [sender.py:send_request():406] send_request: stop_status +2024-06-08 19:03:34,353 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:03:34,434 DEBUG SenderThread:30410 [sender.py:send():379] send: telemetry +2024-06-08 19:03:34,743 INFO wandb-upload_0:30410 [upload_job.py:push():130] Uploaded file /tmp/tmpggzsy1awwandb/624uquwx-wandb-metadata.json +2024-06-08 19:03:34,991 INFO Thread-12 :30410 [dir_watcher.py:_on_file_created():271] file/dir created: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_190333-82mnef5m/files/wandb-metadata.json +2024-06-08 19:03:34,991 INFO Thread-12 :30410 [dir_watcher.py:_on_file_created():271] file/dir created: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_190333-82mnef5m/files/output.log +2024-06-08 19:03:34,991 INFO Thread-12 :30410 [dir_watcher.py:_on_file_created():271] file/dir created: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_190333-82mnef5m/files/requirements.txt +2024-06-08 19:03:35,351 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:03:36,352 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:03:36,991 INFO Thread-12 :30410 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_190333-82mnef5m/files/output.log +2024-06-08 19:03:37,352 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:03:38,354 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:03:39,351 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 19:03:39,355 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:03:40,357 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:03:41,358 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:03:42,359 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:03:43,359 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:03:44,658 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 19:03:44,997 INFO Thread-12 :30410 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_190333-82mnef5m/files/output.log +2024-06-08 19:03:45,767 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:03:46,767 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:03:47,767 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:03:48,767 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:03:49,003 INFO Thread-12 :30410 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_190333-82mnef5m/files/output.log +2024-06-08 19:03:49,350 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: stop_status +2024-06-08 19:03:49,351 DEBUG SenderThread:30410 [sender.py:send_request():406] send_request: stop_status +2024-06-08 19:03:49,767 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:03:50,507 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 19:03:50,767 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:03:51,767 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:03:52,768 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:03:53,768 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:03:54,768 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:03:55,767 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 19:03:55,768 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:03:56,768 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:03:57,768 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:03:58,768 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:03:59,768 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:04:00,767 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 19:04:00,768 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:04:01,768 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:04:02,768 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:04:03,768 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:04:04,350 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: stop_status +2024-06-08 19:04:04,351 DEBUG SenderThread:30410 [sender.py:send_request():406] send_request: stop_status +2024-06-08 19:04:04,769 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:04:05,768 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:04:06,772 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 19:04:06,772 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:04:07,029 INFO Thread-12 :30410 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_190333-82mnef5m/files/config.yaml +2024-06-08 19:04:07,772 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:04:08,772 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:04:09,772 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:04:10,773 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:04:11,355 INFO Thread-12 :30410 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_190333-82mnef5m/files/output.log +2024-06-08 19:04:11,773 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:04:12,699 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 19:04:12,773 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:04:13,357 INFO Thread-12 :30410 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_190333-82mnef5m/files/output.log +2024-06-08 19:04:13,773 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:04:14,773 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:04:15,773 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:04:16,773 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:04:17,767 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 19:04:17,773 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:04:18,773 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:04:19,350 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: stop_status +2024-06-08 19:04:19,350 DEBUG SenderThread:30410 [sender.py:send_request():406] send_request: stop_status +2024-06-08 19:04:19,773 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:04:20,773 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:04:21,774 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:04:22,774 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:04:22,942 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 19:04:23,369 INFO Thread-12 :30410 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_190333-82mnef5m/files/output.log +2024-06-08 19:04:23,774 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:04:24,774 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:04:25,774 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:04:26,774 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:04:27,376 INFO Thread-12 :30410 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_190333-82mnef5m/files/output.log +2024-06-08 19:04:27,774 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:04:28,760 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 19:04:28,806 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:04:29,378 INFO Thread-12 :30410 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_190333-82mnef5m/files/output.log +2024-06-08 19:04:29,774 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:04:30,774 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:04:31,381 INFO Thread-12 :30410 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_190333-82mnef5m/files/output.log +2024-06-08 19:04:31,774 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:04:32,774 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:04:33,778 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:04:34,075 DEBUG SystemMonitor:30410 [system_monitor.py:_start():172] Starting system metrics aggregation loop +2024-06-08 19:04:34,077 DEBUG SenderThread:30410 [sender.py:send():379] send: stats +2024-06-08 19:04:34,078 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 19:04:34,351 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: stop_status +2024-06-08 19:04:34,351 DEBUG SenderThread:30410 [sender.py:send_request():406] send_request: stop_status +2024-06-08 19:04:34,778 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:04:35,483 INFO Thread-12 :30410 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_190333-82mnef5m/files/output.log +2024-06-08 19:04:35,778 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:04:36,778 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:04:37,527 INFO Thread-12 :30410 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_190333-82mnef5m/files/output.log +2024-06-08 19:04:37,778 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:04:38,778 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:04:39,768 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 19:04:39,778 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:04:40,778 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:04:40,847 INFO Thread-12 :30410 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_190333-82mnef5m/files/output.log +2024-06-08 19:04:41,778 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:04:42,784 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:04:43,200 INFO Thread-12 :30410 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_190333-82mnef5m/files/output.log +2024-06-08 19:04:43,789 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:04:44,794 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:04:44,814 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 19:04:45,362 INFO Thread-12 :30410 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_190333-82mnef5m/files/output.log +2024-06-08 19:04:45,794 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:04:46,607 INFO Thread-12 :30410 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_190333-82mnef5m/files/output.log +2024-06-08 19:04:46,799 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:04:47,805 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:04:48,810 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:04:48,915 INFO Thread-12 :30410 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_190333-82mnef5m/files/output.log +2024-06-08 19:04:49,356 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: stop_status +2024-06-08 19:04:49,356 DEBUG SenderThread:30410 [sender.py:send_request():406] send_request: stop_status +2024-06-08 19:04:49,815 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:04:49,899 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 19:04:50,854 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:04:51,172 INFO Thread-12 :30410 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_190333-82mnef5m/files/output.log +2024-06-08 19:04:51,825 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:04:52,831 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:04:53,539 INFO Thread-12 :30410 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_190333-82mnef5m/files/output.log +2024-06-08 19:04:53,831 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:04:54,667 INFO Thread-12 :30410 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_190333-82mnef5m/files/output.log +2024-06-08 19:04:54,831 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:04:55,312 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 19:04:55,831 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:04:56,831 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:04:57,831 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:04:58,831 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:04:59,831 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:05:00,773 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 19:05:00,831 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:05:01,831 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:05:02,831 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:05:03,831 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:05:04,080 DEBUG SenderThread:30410 [sender.py:send():379] send: stats +2024-06-08 19:05:04,356 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: stop_status +2024-06-08 19:05:04,356 DEBUG SenderThread:30410 [sender.py:send_request():406] send_request: stop_status +2024-06-08 19:05:04,831 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:05:05,773 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 19:05:05,832 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:05:06,832 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:05:07,832 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:05:08,560 INFO Thread-12 :30410 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_190333-82mnef5m/files/output.log +2024-06-08 19:05:08,893 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:05:09,893 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:05:10,687 INFO Thread-12 :30410 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_190333-82mnef5m/files/output.log +2024-06-08 19:05:10,773 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 19:05:10,893 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:05:11,700 INFO Thread-12 :30410 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_190333-82mnef5m/files/output.log +2024-06-08 19:05:11,894 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:05:12,894 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:05:13,894 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:05:14,894 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:05:15,773 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 19:05:15,894 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:05:16,894 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:05:17,894 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:05:18,894 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:05:19,356 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: stop_status +2024-06-08 19:05:19,356 DEBUG SenderThread:30410 [sender.py:send_request():406] send_request: stop_status +2024-06-08 19:05:19,894 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:05:20,894 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:05:21,774 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 19:05:21,894 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:05:22,895 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:05:23,895 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:05:24,895 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:05:25,895 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:05:26,895 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:05:27,774 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 19:05:27,895 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:05:28,895 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:05:29,895 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:05:30,895 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:05:31,895 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:05:32,774 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 19:05:32,895 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:05:33,896 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:05:34,082 DEBUG SenderThread:30410 [sender.py:send():379] send: stats +2024-06-08 19:05:34,356 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: stop_status +2024-06-08 19:05:34,356 DEBUG SenderThread:30410 [sender.py:send_request():406] send_request: stop_status +2024-06-08 19:05:34,896 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:05:35,896 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:05:36,896 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:05:37,774 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 19:05:37,896 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:05:38,896 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:05:39,896 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:05:40,896 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:05:41,896 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:05:42,896 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:05:43,774 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 19:05:43,897 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:05:44,897 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:05:45,897 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:05:46,897 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:05:47,897 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:05:48,897 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:05:49,356 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: stop_status +2024-06-08 19:05:49,356 DEBUG SenderThread:30410 [sender.py:send_request():406] send_request: stop_status +2024-06-08 19:05:49,477 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 19:05:49,897 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:05:50,897 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:05:51,897 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:05:52,897 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:05:53,897 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:05:54,775 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 19:05:54,898 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:05:55,898 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:05:56,898 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:05:57,898 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:05:58,898 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:05:59,898 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:06:00,774 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 19:06:00,898 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:06:01,898 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:06:02,898 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:06:03,898 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:06:04,084 DEBUG SenderThread:30410 [sender.py:send():379] send: stats +2024-06-08 19:06:04,356 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: stop_status +2024-06-08 19:06:04,356 DEBUG SenderThread:30410 [sender.py:send_request():406] send_request: stop_status +2024-06-08 19:06:04,898 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:06:05,774 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 19:06:05,898 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:06:06,899 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:06:07,899 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:06:08,899 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:06:09,899 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:06:10,775 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 19:06:10,899 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:06:11,899 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:06:12,899 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:06:13,899 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:06:14,899 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:06:15,775 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 19:06:15,899 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:06:16,899 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:06:17,900 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:06:18,900 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:06:19,356 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: stop_status +2024-06-08 19:06:19,357 DEBUG SenderThread:30410 [sender.py:send_request():406] send_request: stop_status +2024-06-08 19:06:19,900 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:06:20,775 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 19:06:20,900 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:06:21,900 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:06:22,900 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:06:23,900 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:06:24,900 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:06:25,775 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 19:06:25,900 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:06:26,900 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:06:27,900 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:06:28,900 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:06:29,901 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:06:30,775 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 19:06:30,901 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:06:31,901 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:06:32,901 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:06:33,901 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:06:34,086 DEBUG SenderThread:30410 [sender.py:send():379] send: stats +2024-06-08 19:06:34,356 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: stop_status +2024-06-08 19:06:34,357 DEBUG SenderThread:30410 [sender.py:send_request():406] send_request: stop_status +2024-06-08 19:06:34,901 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:06:35,901 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:06:36,775 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 19:06:36,901 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:06:37,901 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:06:38,901 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:06:39,902 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:06:40,902 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:06:41,775 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 19:06:41,902 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:06:42,902 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:06:43,902 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:06:44,902 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:06:44,972 INFO Thread-12 :30410 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_190333-82mnef5m/files/output.log +2024-06-08 19:06:45,902 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:06:46,775 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 19:06:46,902 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:06:47,902 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:06:48,902 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:06:49,356 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: stop_status +2024-06-08 19:06:49,357 DEBUG SenderThread:30410 [sender.py:send_request():406] send_request: stop_status +2024-06-08 19:06:49,902 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:06:50,902 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:06:51,903 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:06:52,775 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 19:06:52,903 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:06:53,903 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:06:54,903 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:06:55,903 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:06:56,903 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:06:57,776 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 19:06:57,903 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:06:58,903 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:06:59,903 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:07:00,903 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:07:01,459 INFO Thread-12 :30410 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_190333-82mnef5m/files/output.log +2024-06-08 19:07:01,904 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:07:02,904 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:07:03,776 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 19:07:03,904 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:07:04,088 DEBUG SenderThread:30410 [sender.py:send():379] send: stats +2024-06-08 19:07:04,357 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: stop_status +2024-06-08 19:07:04,357 DEBUG SenderThread:30410 [sender.py:send_request():406] send_request: stop_status +2024-06-08 19:07:04,904 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:07:05,904 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:07:06,904 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:07:07,904 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:07:08,776 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 19:07:08,904 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:07:09,904 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:07:10,904 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:07:11,092 INFO Thread-12 :30410 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_190333-82mnef5m/files/output.log +2024-06-08 19:07:11,905 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:07:12,905 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:07:13,905 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:07:14,776 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 19:07:14,905 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:07:15,905 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:07:16,905 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:07:17,905 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:07:18,905 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:07:19,357 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: stop_status +2024-06-08 19:07:19,357 DEBUG SenderThread:30410 [sender.py:send_request():406] send_request: stop_status +2024-06-08 19:07:19,905 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:07:20,445 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 19:07:20,905 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:07:21,905 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:07:22,906 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:07:23,906 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:07:24,906 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:07:25,511 INFO Thread-12 :30410 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_190333-82mnef5m/files/output.log +2024-06-08 19:07:25,776 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 19:07:25,906 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:07:26,906 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:07:27,906 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:07:28,906 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:07:29,906 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:07:30,906 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:07:31,776 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 19:07:31,906 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:07:32,906 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:07:33,906 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:07:34,090 DEBUG SenderThread:30410 [sender.py:send():379] send: stats +2024-06-08 19:07:34,357 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: stop_status +2024-06-08 19:07:34,357 DEBUG SenderThread:30410 [sender.py:send_request():406] send_request: stop_status +2024-06-08 19:07:34,906 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:07:35,906 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:07:36,776 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 19:07:36,907 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:07:37,156 INFO Thread-12 :30410 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_190333-82mnef5m/files/output.log +2024-06-08 19:07:37,907 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:07:38,907 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:07:39,907 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:07:40,907 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:07:41,907 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:07:42,776 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 19:07:42,907 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:07:43,907 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:07:44,907 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:07:45,907 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:07:46,908 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:07:47,776 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 19:07:47,908 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:07:48,908 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:07:49,357 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: stop_status +2024-06-08 19:07:49,357 DEBUG SenderThread:30410 [sender.py:send_request():406] send_request: stop_status +2024-06-08 19:07:49,908 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:07:50,908 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:07:51,459 INFO Thread-12 :30410 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_190333-82mnef5m/files/output.log +2024-06-08 19:07:51,908 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:07:52,908 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:07:53,776 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 19:07:53,908 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:07:54,908 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:07:55,908 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:07:56,909 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:07:57,909 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:07:58,909 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:07:59,729 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 19:07:59,909 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:08:00,909 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:08:01,000 INFO Thread-12 :30410 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_190333-82mnef5m/files/output.log +2024-06-08 19:08:01,909 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:08:02,909 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:08:03,909 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:08:04,092 DEBUG SenderThread:30410 [sender.py:send():379] send: stats +2024-06-08 19:08:04,357 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: stop_status +2024-06-08 19:08:04,357 DEBUG SenderThread:30410 [sender.py:send_request():406] send_request: stop_status +2024-06-08 19:08:04,909 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:08:05,504 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 19:08:05,909 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:08:06,909 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:08:07,910 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:08:08,910 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:08:09,910 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:08:10,777 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 19:08:10,910 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:08:11,016 INFO Thread-12 :30410 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_190333-82mnef5m/files/output.log +2024-06-08 19:08:11,910 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:08:12,910 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:08:13,910 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:08:14,910 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:08:15,777 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 19:08:15,910 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:08:16,910 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:08:17,910 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:08:18,911 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:08:19,357 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: stop_status +2024-06-08 19:08:19,357 DEBUG SenderThread:30410 [sender.py:send_request():406] send_request: stop_status +2024-06-08 19:08:19,911 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:08:20,911 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:08:21,435 INFO Thread-12 :30410 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_190333-82mnef5m/files/output.log +2024-06-08 19:08:21,777 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 19:08:21,911 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:08:22,911 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:08:23,911 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:08:24,911 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:08:25,911 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:08:26,777 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 19:08:26,911 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:08:27,911 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:08:28,911 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:08:29,912 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:08:30,912 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:08:31,777 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 19:08:31,912 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:08:32,912 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:08:33,305 INFO Thread-12 :30410 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_190333-82mnef5m/files/output.log +2024-06-08 19:08:33,912 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:08:34,095 DEBUG SenderThread:30410 [sender.py:send():379] send: stats +2024-06-08 19:08:34,357 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: stop_status +2024-06-08 19:08:34,357 DEBUG SenderThread:30410 [sender.py:send_request():406] send_request: stop_status +2024-06-08 19:08:34,912 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:08:35,912 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:08:36,912 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:08:37,777 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 19:08:37,912 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:08:38,912 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:08:39,912 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:08:40,912 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:08:41,913 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:08:42,777 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 19:08:42,913 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:08:43,913 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:08:44,913 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:08:45,012 INFO Thread-12 :30410 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_190333-82mnef5m/files/output.log +2024-06-08 19:08:45,930 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:08:46,913 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:08:47,777 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 19:08:47,913 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:08:48,913 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:08:49,357 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: stop_status +2024-06-08 19:08:49,357 DEBUG SenderThread:30410 [sender.py:send_request():406] send_request: stop_status +2024-06-08 19:08:49,913 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:08:50,913 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:08:51,913 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:08:52,914 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:08:53,778 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 19:08:53,914 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:08:54,914 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:08:55,914 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:08:56,914 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:08:57,914 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:08:58,778 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 19:08:58,914 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:08:59,914 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:09:00,914 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:09:01,460 INFO Thread-12 :30410 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_190333-82mnef5m/files/output.log +2024-06-08 19:09:01,914 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:09:02,915 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:09:03,915 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:09:04,096 DEBUG SenderThread:30410 [sender.py:send():379] send: stats +2024-06-08 19:09:04,099 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 19:09:04,357 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: stop_status +2024-06-08 19:09:04,358 DEBUG SenderThread:30410 [sender.py:send_request():406] send_request: stop_status +2024-06-08 19:09:04,915 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:09:05,915 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:09:06,915 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:09:07,396 INFO Thread-12 :30410 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_190333-82mnef5m/files/output.log +2024-06-08 19:09:07,915 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:09:08,915 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:09:09,915 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:09:09,956 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 19:09:10,915 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:09:11,915 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:09:12,915 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:09:13,915 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:09:14,916 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:09:15,778 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 19:09:15,916 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:09:16,916 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:09:17,916 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:09:18,916 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:09:19,357 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: stop_status +2024-06-08 19:09:19,358 DEBUG SenderThread:30410 [sender.py:send_request():406] send_request: stop_status +2024-06-08 19:09:19,916 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:09:20,778 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 19:09:20,916 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:09:21,431 INFO Thread-12 :30410 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_190333-82mnef5m/files/output.log +2024-06-08 19:09:21,916 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:09:22,916 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:09:23,916 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:09:24,916 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:09:25,778 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 19:09:25,916 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:09:26,917 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:09:27,917 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:09:28,922 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:09:29,400 INFO Thread-12 :30410 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_190333-82mnef5m/files/output.log +2024-06-08 19:09:29,922 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:09:30,778 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 19:09:30,922 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:09:31,922 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:09:32,922 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:09:33,923 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:09:34,102 DEBUG SenderThread:30410 [sender.py:send():379] send: stats +2024-06-08 19:09:34,357 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: stop_status +2024-06-08 19:09:34,358 DEBUG SenderThread:30410 [sender.py:send_request():406] send_request: stop_status +2024-06-08 19:09:34,923 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:09:35,778 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 19:09:35,923 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:09:36,923 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:09:37,923 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:09:38,923 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:09:39,923 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:09:40,923 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:09:41,284 INFO Thread-12 :30410 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_190333-82mnef5m/files/output.log +2024-06-08 19:09:41,778 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 19:09:41,923 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:09:42,923 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:09:43,923 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:09:44,924 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:09:45,924 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:09:46,778 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 19:09:46,924 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:09:47,924 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:09:48,924 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:09:49,357 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: stop_status +2024-06-08 19:09:49,358 DEBUG SenderThread:30410 [sender.py:send_request():406] send_request: stop_status +2024-06-08 19:09:49,495 INFO Thread-12 :30410 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_190333-82mnef5m/files/output.log +2024-06-08 19:09:49,924 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:09:50,924 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:09:51,924 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:09:52,779 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 19:09:52,924 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:09:53,924 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:09:54,924 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:09:55,925 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:09:56,925 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:09:57,779 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 19:09:57,925 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:09:58,925 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:09:59,925 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:10:00,925 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:10:01,435 INFO Thread-12 :30410 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_190333-82mnef5m/files/output.log +2024-06-08 19:10:01,925 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:10:02,925 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:10:03,779 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 19:10:03,925 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:10:04,104 DEBUG SenderThread:30410 [sender.py:send():379] send: stats +2024-06-08 19:10:04,358 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: stop_status +2024-06-08 19:10:04,358 DEBUG SenderThread:30410 [sender.py:send_request():406] send_request: stop_status +2024-06-08 19:10:04,925 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:10:05,925 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:10:06,926 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:10:07,926 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:10:08,926 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:10:09,387 INFO Thread-12 :30410 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_190333-82mnef5m/files/output.log +2024-06-08 19:10:09,510 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 19:10:09,926 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:10:10,926 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:10:11,891 INFO Thread-12 :30410 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_190333-82mnef5m/files/output.log +2024-06-08 19:10:11,926 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:10:12,926 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:10:13,926 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:10:14,788 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 19:10:14,926 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:10:15,926 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:10:16,926 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:10:17,927 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:10:18,927 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:10:19,358 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: stop_status +2024-06-08 19:10:19,358 DEBUG SenderThread:30410 [sender.py:send_request():406] send_request: stop_status +2024-06-08 19:10:19,927 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:10:20,502 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 19:10:20,927 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:10:21,415 INFO Thread-12 :30410 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_190333-82mnef5m/files/output.log +2024-06-08 19:10:21,927 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:10:22,927 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:10:23,927 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:10:24,927 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:10:25,779 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 19:10:25,927 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:10:26,927 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:10:27,927 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:10:28,928 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:10:29,928 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:10:30,779 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 19:10:30,904 INFO Thread-12 :30410 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_190333-82mnef5m/files/output.log +2024-06-08 19:10:30,928 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:10:31,928 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:10:32,928 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:10:33,928 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:10:34,107 DEBUG SenderThread:30410 [sender.py:send():379] send: stats +2024-06-08 19:10:34,358 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: stop_status +2024-06-08 19:10:34,358 DEBUG SenderThread:30410 [sender.py:send_request():406] send_request: stop_status +2024-06-08 19:10:34,928 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:10:35,928 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:10:36,779 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 19:10:36,928 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:10:37,929 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:10:38,929 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:10:39,929 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:10:40,929 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:10:41,547 INFO Thread-12 :30410 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_190333-82mnef5m/files/output.log +2024-06-08 19:10:41,929 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:10:42,779 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 19:10:42,929 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:10:43,929 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:10:44,929 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:10:45,080 INFO Thread-12 :30410 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_190333-82mnef5m/files/output.log +2024-06-08 19:10:45,929 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:10:46,929 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:10:47,328 INFO Thread-12 :30410 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_190333-82mnef5m/files/output.log +2024-06-08 19:10:47,780 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 19:10:47,929 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:10:48,930 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:10:49,358 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: stop_status +2024-06-08 19:10:49,358 DEBUG SenderThread:30410 [sender.py:send_request():406] send_request: stop_status +2024-06-08 19:10:49,930 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:10:50,930 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:10:51,930 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:10:52,780 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 19:10:52,930 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:10:53,930 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:10:54,930 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:10:55,930 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:10:56,930 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:10:57,930 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:10:58,780 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 19:10:58,931 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:10:59,931 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:11:00,935 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:11:01,627 INFO Thread-12 :30410 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_190333-82mnef5m/files/output.log +2024-06-08 19:11:01,936 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:11:02,936 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:11:03,936 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:11:04,108 DEBUG SenderThread:30410 [sender.py:send():379] send: stats +2024-06-08 19:11:04,111 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 19:11:04,358 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: stop_status +2024-06-08 19:11:04,358 DEBUG SenderThread:30410 [sender.py:send_request():406] send_request: stop_status +2024-06-08 19:11:04,936 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:11:05,936 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:11:06,936 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:11:07,936 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:11:08,936 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:11:09,780 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 19:11:09,936 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:11:10,936 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:11:11,937 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:11:12,937 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:11:13,112 INFO Thread-12 :30410 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_190333-82mnef5m/files/output.log +2024-06-08 19:11:13,937 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:11:14,937 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:11:15,267 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 19:11:15,937 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:11:16,937 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:11:17,581 INFO Thread-12 :30410 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_190333-82mnef5m/files/output.log +2024-06-08 19:11:17,937 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:11:18,937 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:11:19,358 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: stop_status +2024-06-08 19:11:19,358 DEBUG SenderThread:30410 [sender.py:send_request():406] send_request: stop_status +2024-06-08 19:11:19,937 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:11:20,453 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 19:11:20,937 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:11:21,937 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:11:22,938 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:11:23,938 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:11:24,938 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:11:25,780 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 19:11:25,938 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:11:26,938 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:11:27,938 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:11:28,938 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:11:29,938 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:11:30,780 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 19:11:30,938 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:11:31,891 INFO Thread-12 :30410 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_190333-82mnef5m/files/output.log +2024-06-08 19:11:31,938 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:11:32,938 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:11:33,229 INFO Thread-12 :30410 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_190333-82mnef5m/files/output.log +2024-06-08 19:11:33,939 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:11:34,111 DEBUG SenderThread:30410 [sender.py:send():379] send: stats +2024-06-08 19:11:34,358 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: stop_status +2024-06-08 19:11:34,358 DEBUG SenderThread:30410 [sender.py:send_request():406] send_request: stop_status +2024-06-08 19:11:34,939 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:11:35,539 INFO Thread-12 :30410 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_190333-82mnef5m/files/output.log +2024-06-08 19:11:35,781 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 19:11:35,939 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:11:36,939 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:11:37,939 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:11:38,939 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:11:39,939 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:11:40,781 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 19:11:40,940 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:11:41,941 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:11:42,941 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:11:43,941 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:11:44,941 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:11:45,780 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 19:11:45,941 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:11:46,941 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:11:47,941 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:11:48,941 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:11:49,358 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: stop_status +2024-06-08 19:11:49,358 DEBUG SenderThread:30410 [sender.py:send_request():406] send_request: stop_status +2024-06-08 19:11:49,941 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:11:50,941 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:11:51,781 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 19:11:51,941 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:11:52,123 INFO Thread-12 :30410 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_190333-82mnef5m/files/output.log +2024-06-08 19:11:52,942 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:11:53,942 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:11:54,942 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:11:55,942 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:11:56,781 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 19:11:56,942 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:11:57,942 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:11:58,942 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:11:59,942 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:12:00,942 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:12:01,942 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:12:02,268 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 19:12:02,942 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:12:03,667 INFO Thread-12 :30410 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_190333-82mnef5m/files/output.log +2024-06-08 19:12:03,943 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:12:04,113 DEBUG SenderThread:30410 [sender.py:send():379] send: stats +2024-06-08 19:12:04,358 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: stop_status +2024-06-08 19:12:04,359 DEBUG SenderThread:30410 [sender.py:send_request():406] send_request: stop_status +2024-06-08 19:12:04,942 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:12:05,943 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:12:06,943 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:12:07,047 INFO Thread-12 :30410 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_190333-82mnef5m/files/output.log +2024-06-08 19:12:07,781 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 19:12:07,943 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:12:08,943 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:12:09,943 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:12:10,943 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:12:11,943 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:12:12,943 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:12:13,781 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 19:12:13,943 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:12:14,944 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:12:15,944 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:12:16,944 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:12:17,944 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:12:18,781 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 19:12:18,944 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:12:19,358 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: stop_status +2024-06-08 19:12:19,359 DEBUG SenderThread:30410 [sender.py:send_request():406] send_request: stop_status +2024-06-08 19:12:19,944 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:12:20,944 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:12:21,044 INFO Thread-12 :30410 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_190333-82mnef5m/files/output.log +2024-06-08 19:12:21,944 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:12:22,944 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:12:23,316 INFO Thread-12 :30410 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_190333-82mnef5m/files/output.log +2024-06-08 19:12:23,944 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:12:24,781 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 19:12:24,944 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:12:25,944 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:12:26,945 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:12:27,945 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:12:28,945 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:12:29,782 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 19:12:29,945 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:12:30,945 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:12:31,945 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:12:32,945 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:12:33,945 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:12:33,947 INFO Thread-12 :30410 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_190333-82mnef5m/files/output.log +2024-06-08 19:12:34,115 DEBUG SenderThread:30410 [sender.py:send():379] send: stats +2024-06-08 19:12:34,359 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: stop_status +2024-06-08 19:12:34,359 DEBUG SenderThread:30410 [sender.py:send_request():406] send_request: stop_status +2024-06-08 19:12:34,850 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 19:12:34,945 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:12:35,945 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:12:36,951 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:12:37,380 INFO Thread-12 :30410 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_190333-82mnef5m/files/output.log +2024-06-08 19:12:37,946 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:12:38,946 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:12:39,946 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:12:40,782 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 19:12:40,946 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:12:41,946 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:12:42,946 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:12:43,946 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:12:44,946 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:12:45,782 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 19:12:45,946 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:12:46,946 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:12:47,946 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:12:48,947 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:12:49,358 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: stop_status +2024-06-08 19:12:49,359 DEBUG SenderThread:30410 [sender.py:send_request():406] send_request: stop_status +2024-06-08 19:12:49,947 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:12:50,947 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:12:51,409 INFO Thread-12 :30410 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_190333-82mnef5m/files/output.log +2024-06-08 19:12:51,782 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 19:12:51,947 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:12:52,947 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:12:53,947 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:12:54,947 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:12:55,947 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:12:56,782 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 19:12:56,947 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:12:57,947 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:12:58,947 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:12:59,948 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:13:00,948 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:13:01,948 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:13:02,270 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 19:13:02,948 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:13:03,948 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:13:03,955 INFO Thread-12 :30410 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_190333-82mnef5m/files/output.log +2024-06-08 19:13:04,117 DEBUG SenderThread:30410 [sender.py:send():379] send: stats +2024-06-08 19:13:04,359 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: stop_status +2024-06-08 19:13:04,359 DEBUG SenderThread:30410 [sender.py:send_request():406] send_request: stop_status +2024-06-08 19:13:04,948 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:13:05,153 INFO Thread-12 :30410 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_190333-82mnef5m/files/output.log +2024-06-08 19:13:05,958 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:13:06,958 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:13:07,782 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 19:13:07,958 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:13:08,958 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:13:09,958 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:13:10,958 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:13:11,959 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:13:12,959 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:13:13,782 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 19:13:13,959 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:13:14,959 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:13:15,959 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:13:16,959 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:13:17,959 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:13:18,784 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 19:13:18,959 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:13:19,200 INFO Thread-12 :30410 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_190333-82mnef5m/files/output.log +2024-06-08 19:13:19,359 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: stop_status +2024-06-08 19:13:19,359 DEBUG SenderThread:30410 [sender.py:send_request():406] send_request: stop_status +2024-06-08 19:13:19,959 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:13:20,959 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:13:21,387 INFO Thread-12 :30410 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_190333-82mnef5m/files/output.log +2024-06-08 19:13:21,959 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:13:22,959 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:13:23,960 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:13:24,783 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 19:13:24,960 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:13:25,960 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:13:26,960 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:13:27,960 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:13:28,960 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:13:29,960 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:13:30,168 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 19:13:30,960 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:13:31,931 INFO Thread-12 :30410 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_190333-82mnef5m/files/output.log +2024-06-08 19:13:31,960 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:13:32,960 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:13:33,961 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:13:34,120 DEBUG SenderThread:30410 [sender.py:send():379] send: stats +2024-06-08 19:13:34,359 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: stop_status +2024-06-08 19:13:34,359 DEBUG SenderThread:30410 [sender.py:send_request():406] send_request: stop_status +2024-06-08 19:13:34,978 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:13:35,444 INFO Thread-12 :30410 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_190333-82mnef5m/files/output.log +2024-06-08 19:13:35,499 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 19:13:35,961 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:13:36,961 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:13:37,961 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:13:38,961 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:13:39,961 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:13:40,783 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 19:13:40,961 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:13:41,961 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:13:42,961 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:13:43,961 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:13:44,961 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:13:45,783 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 19:13:45,962 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:13:46,962 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:13:47,962 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:13:48,962 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:13:49,359 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: stop_status +2024-06-08 19:13:49,359 DEBUG SenderThread:30410 [sender.py:send_request():406] send_request: stop_status +2024-06-08 19:13:49,427 INFO Thread-12 :30410 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_190333-82mnef5m/files/output.log +2024-06-08 19:13:49,962 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:13:50,783 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 19:13:50,962 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:13:51,962 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:13:52,962 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:13:53,962 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:13:54,962 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:13:55,783 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 19:13:55,962 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:13:56,963 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:13:57,963 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:13:58,963 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:13:59,963 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:14:00,783 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 19:14:00,963 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:14:01,963 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:14:02,291 INFO Thread-12 :30410 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_190333-82mnef5m/files/output.log +2024-06-08 19:14:02,963 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:14:03,419 INFO Thread-12 :30410 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_190333-82mnef5m/files/output.log +2024-06-08 19:14:03,963 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:14:04,122 DEBUG SenderThread:30410 [sender.py:send():379] send: stats +2024-06-08 19:14:04,362 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: stop_status +2024-06-08 19:14:04,362 DEBUG SenderThread:30410 [sender.py:send_request():406] send_request: stop_status +2024-06-08 19:14:04,963 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:14:05,665 INFO Thread-12 :30410 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_190333-82mnef5m/files/output.log +2024-06-08 19:14:05,783 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 19:14:05,963 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:14:06,963 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:14:07,964 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:14:08,964 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:14:09,964 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:14:10,783 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 19:14:10,964 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:14:11,964 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:14:12,964 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:14:13,964 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:14:14,964 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:14:15,784 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 19:14:15,964 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:14:16,964 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:14:17,671 INFO Thread-12 :30410 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_190333-82mnef5m/files/output.log +2024-06-08 19:14:17,964 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:14:18,965 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:14:19,362 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: stop_status +2024-06-08 19:14:19,362 DEBUG SenderThread:30410 [sender.py:send_request():406] send_request: stop_status +2024-06-08 19:14:19,965 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:14:20,965 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:14:21,784 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 19:14:21,965 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:14:22,965 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:14:23,965 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:14:24,965 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:14:25,965 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:14:26,965 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:14:27,784 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 19:14:27,965 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:14:28,965 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:14:29,966 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:14:30,966 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:14:31,403 INFO Thread-12 :30410 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_190333-82mnef5m/files/output.log +2024-06-08 19:14:31,966 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:14:32,966 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:14:33,540 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 19:14:33,592 INFO Thread-12 :30410 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_190333-82mnef5m/files/output.log +2024-06-08 19:14:33,966 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:14:34,124 DEBUG SenderThread:30410 [sender.py:send():379] send: stats +2024-06-08 19:14:34,362 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: stop_status +2024-06-08 19:14:34,362 DEBUG SenderThread:30410 [sender.py:send_request():406] send_request: stop_status +2024-06-08 19:14:34,966 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:14:35,966 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:14:36,966 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:14:37,966 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:14:38,784 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 19:14:38,966 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:14:39,967 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:14:40,967 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:14:41,967 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:14:42,967 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:14:43,967 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:14:44,784 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 19:14:44,967 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:14:45,967 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:14:46,967 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:14:47,967 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:14:48,967 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:14:49,362 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: stop_status +2024-06-08 19:14:49,362 DEBUG SenderThread:30410 [sender.py:send_request():406] send_request: stop_status +2024-06-08 19:14:49,967 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:14:50,115 INFO Thread-12 :30410 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_190333-82mnef5m/files/output.log +2024-06-08 19:14:50,487 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 19:14:50,967 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:14:51,968 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:14:52,968 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:14:53,968 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:14:54,968 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:14:55,784 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 19:14:55,968 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:14:56,968 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:14:57,968 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:14:58,968 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:14:59,968 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:15:00,968 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:15:01,707 INFO Thread-12 :30410 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_190333-82mnef5m/files/output.log +2024-06-08 19:15:01,784 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 19:15:01,969 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:15:02,969 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:15:03,969 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:15:04,127 DEBUG SenderThread:30410 [sender.py:send():379] send: stats +2024-06-08 19:15:04,362 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: stop_status +2024-06-08 19:15:04,362 DEBUG SenderThread:30410 [sender.py:send_request():406] send_request: stop_status +2024-06-08 19:15:04,969 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:15:05,269 INFO Thread-12 :30410 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_190333-82mnef5m/files/output.log +2024-06-08 19:15:05,969 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:15:06,784 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 19:15:06,969 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:15:07,969 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:15:08,969 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:15:09,969 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:15:10,969 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:15:11,969 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:15:12,785 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 19:15:12,970 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:15:13,970 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:15:14,970 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:15:15,970 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:15:16,970 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:15:17,970 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:15:18,630 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 19:15:18,970 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:15:19,362 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: stop_status +2024-06-08 19:15:19,363 DEBUG SenderThread:30410 [sender.py:send_request():406] send_request: stop_status +2024-06-08 19:15:19,571 INFO Thread-12 :30410 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_190333-82mnef5m/files/output.log +2024-06-08 19:15:19,970 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:15:20,978 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:15:21,675 INFO Thread-12 :30410 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_190333-82mnef5m/files/output.log +2024-06-08 19:15:21,978 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:15:22,978 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:15:23,785 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 19:15:23,978 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:15:24,978 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:15:25,978 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:15:26,979 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:15:27,979 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:15:28,979 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:15:29,173 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 19:15:29,979 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:15:30,979 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:15:31,515 INFO Thread-12 :30410 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_190333-82mnef5m/files/output.log +2024-06-08 19:15:31,979 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:15:32,979 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:15:33,732 INFO Thread-12 :30410 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_190333-82mnef5m/files/output.log +2024-06-08 19:15:33,979 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:15:34,129 DEBUG SenderThread:30410 [sender.py:send():379] send: stats +2024-06-08 19:15:34,362 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: stop_status +2024-06-08 19:15:34,363 DEBUG SenderThread:30410 [sender.py:send_request():406] send_request: stop_status +2024-06-08 19:15:34,443 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 19:15:34,979 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:15:35,895 INFO Thread-12 :30410 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_190333-82mnef5m/files/output.log +2024-06-08 19:15:35,979 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:15:36,979 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:15:37,979 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:15:38,980 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:15:39,785 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 19:15:39,980 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:15:40,980 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:15:41,980 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:15:42,980 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:15:43,980 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:15:44,980 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:15:45,785 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 19:15:45,980 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:15:46,980 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:15:47,980 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:15:48,980 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:15:49,363 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: stop_status +2024-06-08 19:15:49,363 DEBUG SenderThread:30410 [sender.py:send_request():406] send_request: stop_status +2024-06-08 19:15:49,981 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:15:50,311 INFO Thread-12 :30410 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_190333-82mnef5m/files/output.log +2024-06-08 19:15:50,785 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 19:15:50,981 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:15:51,981 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:15:52,981 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:15:53,981 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:15:54,981 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:15:55,785 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 19:15:55,981 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:15:56,981 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:15:57,981 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:15:58,981 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:15:59,982 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:16:00,785 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 19:16:00,982 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:16:01,657 INFO Thread-12 :30410 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_190333-82mnef5m/files/output.log +2024-06-08 19:16:01,982 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:16:02,982 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:16:03,911 INFO Thread-12 :30410 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_190333-82mnef5m/files/output.log +2024-06-08 19:16:03,982 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:16:04,131 DEBUG SenderThread:30410 [sender.py:send():379] send: stats +2024-06-08 19:16:04,363 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: stop_status +2024-06-08 19:16:04,363 DEBUG SenderThread:30410 [sender.py:send_request():406] send_request: stop_status +2024-06-08 19:16:04,982 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:16:05,786 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 19:16:05,982 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:16:06,982 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:16:07,982 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:16:08,982 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:16:09,982 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:16:10,786 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 19:16:10,982 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:16:11,983 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:16:12,983 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:16:13,837 INFO Thread-12 :30410 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_190333-82mnef5m/files/output.log +2024-06-08 19:16:13,983 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:16:14,983 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:16:15,786 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 19:16:15,983 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:16:16,339 INFO Thread-12 :30410 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_190333-82mnef5m/files/output.log +2024-06-08 19:16:16,983 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:16:17,703 INFO Thread-12 :30410 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_190333-82mnef5m/files/output.log +2024-06-08 19:16:17,983 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:16:18,983 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:16:19,363 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: stop_status +2024-06-08 19:16:19,363 DEBUG SenderThread:30410 [sender.py:send_request():406] send_request: stop_status +2024-06-08 19:16:19,983 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:16:20,786 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 19:16:20,983 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:16:21,983 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:16:22,984 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:16:23,984 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:16:24,984 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:16:25,786 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 19:16:25,984 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:16:26,984 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:16:27,984 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:16:28,984 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:16:29,562 INFO Thread-12 :30410 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_190333-82mnef5m/files/output.log +2024-06-08 19:16:29,984 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:16:30,786 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 19:16:30,984 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:16:31,732 INFO Thread-12 :30410 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_190333-82mnef5m/files/output.log +2024-06-08 19:16:31,984 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:16:32,984 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:16:33,985 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:16:34,133 DEBUG SenderThread:30410 [sender.py:send():379] send: stats +2024-06-08 19:16:34,363 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: stop_status +2024-06-08 19:16:34,363 DEBUG SenderThread:30410 [sender.py:send_request():406] send_request: stop_status +2024-06-08 19:16:34,985 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:16:35,985 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:16:36,786 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 19:16:36,985 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:16:37,985 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:16:38,985 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:16:39,985 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:16:40,985 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:16:41,588 INFO Thread-12 :30410 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_190333-82mnef5m/files/output.log +2024-06-08 19:16:41,786 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 19:16:41,985 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:16:42,985 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:16:43,985 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:16:44,032 INFO Thread-12 :30410 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_190333-82mnef5m/files/output.log +2024-06-08 19:16:44,986 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:16:45,986 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:16:46,986 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:16:47,786 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 19:16:47,986 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:16:48,986 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:16:49,363 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: stop_status +2024-06-08 19:16:49,363 DEBUG SenderThread:30410 [sender.py:send_request():406] send_request: stop_status +2024-06-08 19:16:49,986 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:16:50,986 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:16:51,986 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:16:52,986 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:16:53,787 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 19:16:53,986 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:16:54,986 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:16:55,986 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:16:56,987 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:16:57,346 INFO Thread-12 :30410 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_190333-82mnef5m/files/output.log +2024-06-08 19:16:57,987 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:16:58,987 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:16:59,630 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 19:16:59,987 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:17:00,987 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:17:01,987 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:17:02,987 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:17:03,987 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:17:04,135 DEBUG SenderThread:30410 [sender.py:send():379] send: stats +2024-06-08 19:17:04,363 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: stop_status +2024-06-08 19:17:04,363 DEBUG SenderThread:30410 [sender.py:send_request():406] send_request: stop_status +2024-06-08 19:17:04,987 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:17:05,468 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 19:17:05,987 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:17:06,988 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:17:07,988 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:17:08,988 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:17:09,988 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:17:10,787 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 19:17:10,988 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:17:11,296 INFO Thread-12 :30410 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_190333-82mnef5m/files/output.log +2024-06-08 19:17:11,988 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:17:12,988 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:17:13,988 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:17:14,988 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:17:15,787 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 19:17:15,988 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:17:16,988 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:17:17,989 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:17:18,989 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:17:19,363 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: stop_status +2024-06-08 19:17:19,363 DEBUG SenderThread:30410 [sender.py:send_request():406] send_request: stop_status +2024-06-08 19:17:19,989 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:17:20,989 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:17:21,787 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 19:17:21,989 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:17:22,989 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:17:23,989 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:17:24,259 INFO Thread-12 :30410 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_190333-82mnef5m/files/output.log +2024-06-08 19:17:24,989 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:17:25,385 INFO Thread-12 :30410 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_190333-82mnef5m/files/output.log +2024-06-08 19:17:25,989 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:17:26,787 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 19:17:26,989 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:17:27,989 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:17:28,990 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:17:29,990 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:17:30,990 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:17:31,787 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 19:17:31,990 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:17:32,990 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:17:33,990 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:17:34,138 DEBUG SenderThread:30410 [sender.py:send():379] send: stats +2024-06-08 19:17:34,363 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: stop_status +2024-06-08 19:17:34,363 DEBUG SenderThread:30410 [sender.py:send_request():406] send_request: stop_status +2024-06-08 19:17:34,990 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:17:35,990 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:17:36,990 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:17:37,787 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 19:17:37,990 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:17:38,990 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:17:39,991 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:17:40,238 INFO Thread-12 :30410 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_190333-82mnef5m/files/output.log +2024-06-08 19:17:40,991 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:17:41,991 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:17:42,991 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:17:43,787 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 19:17:43,991 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:17:44,991 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:17:45,991 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:17:46,991 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:17:47,991 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:17:48,991 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:17:49,177 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 19:17:49,363 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: stop_status +2024-06-08 19:17:49,364 DEBUG SenderThread:30410 [sender.py:send_request():406] send_request: stop_status +2024-06-08 19:17:49,991 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:17:50,992 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:17:51,992 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:17:52,147 INFO Thread-12 :30410 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_190333-82mnef5m/files/output.log +2024-06-08 19:17:52,992 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:17:53,992 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:17:54,399 INFO Thread-12 :30410 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_190333-82mnef5m/files/output.log +2024-06-08 19:17:54,804 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 19:17:54,992 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:17:55,992 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:17:56,992 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:17:57,992 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:17:58,992 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:17:59,992 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:18:00,788 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 19:18:00,992 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:18:01,993 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:18:02,993 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:18:03,731 INFO Thread-12 :30410 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_190333-82mnef5m/files/output.log +2024-06-08 19:18:03,993 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:18:04,140 DEBUG SenderThread:30410 [sender.py:send():379] send: stats +2024-06-08 19:18:04,363 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: stop_status +2024-06-08 19:18:04,363 DEBUG SenderThread:30410 [sender.py:send_request():406] send_request: stop_status +2024-06-08 19:18:04,993 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:18:05,788 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 19:18:05,993 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:18:06,031 INFO Thread-12 :30410 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_190333-82mnef5m/files/output.log +2024-06-08 19:18:06,993 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:18:07,993 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:18:08,993 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:18:09,993 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:18:10,788 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 19:18:10,993 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:18:11,994 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:18:12,994 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:18:13,994 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:18:14,994 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:18:15,788 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 19:18:15,994 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:18:16,994 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:18:17,994 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:18:18,994 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:18:19,363 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: stop_status +2024-06-08 19:18:19,364 DEBUG SenderThread:30410 [sender.py:send_request():406] send_request: stop_status +2024-06-08 19:18:19,843 INFO Thread-12 :30410 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_190333-82mnef5m/files/output.log +2024-06-08 19:18:19,994 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:18:20,788 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 19:18:20,994 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:18:21,994 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:18:22,994 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:18:23,995 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:18:24,995 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:18:25,788 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 19:18:25,995 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:18:26,995 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:18:27,995 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:18:28,995 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:18:29,995 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:18:30,788 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 19:18:31,002 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:18:31,635 INFO Thread-12 :30410 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_190333-82mnef5m/files/output.log +2024-06-08 19:18:31,995 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:18:32,995 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:18:33,939 INFO Thread-12 :30410 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_190333-82mnef5m/files/output.log +2024-06-08 19:18:33,996 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:18:34,142 DEBUG SenderThread:30410 [sender.py:send():379] send: stats +2024-06-08 19:18:34,364 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: stop_status +2024-06-08 19:18:34,364 DEBUG SenderThread:30410 [sender.py:send_request():406] send_request: stop_status +2024-06-08 19:18:34,996 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:18:35,996 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:18:36,788 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 19:18:36,996 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:18:37,996 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:18:38,996 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:18:39,996 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:18:40,996 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:18:41,996 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:18:42,277 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 19:18:42,996 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:18:43,402 INFO Thread-12 :30410 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_190333-82mnef5m/files/output.log +2024-06-08 19:18:43,996 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:18:44,996 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:18:45,627 INFO Thread-12 :30410 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_190333-82mnef5m/files/output.log +2024-06-08 19:18:45,997 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:18:46,997 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:18:47,788 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 19:18:47,997 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:18:48,997 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:18:49,364 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: stop_status +2024-06-08 19:18:49,364 DEBUG SenderThread:30410 [sender.py:send_request():406] send_request: stop_status +2024-06-08 19:18:49,997 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:18:50,997 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:18:51,997 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:18:52,997 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:18:53,789 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 19:18:53,997 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:18:54,997 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:18:55,998 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:18:56,998 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:18:57,998 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:18:58,998 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:18:58,999 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 19:18:59,788 INFO Thread-12 :30410 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_190333-82mnef5m/files/output.log +2024-06-08 19:18:59,998 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:19:00,998 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:19:01,998 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:19:02,999 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:19:03,999 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:19:04,144 DEBUG SenderThread:30410 [sender.py:send():379] send: stats +2024-06-08 19:19:04,146 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 19:19:04,364 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: stop_status +2024-06-08 19:19:04,364 DEBUG SenderThread:30410 [sender.py:send_request():406] send_request: stop_status +2024-06-08 19:19:04,999 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:19:05,999 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:19:06,999 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:19:07,999 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:19:08,999 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:19:09,179 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 19:19:09,999 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:19:10,999 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:19:11,667 INFO Thread-12 :30410 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_190333-82mnef5m/files/output.log +2024-06-08 19:19:11,999 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:19:12,999 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:19:13,811 INFO Thread-12 :30410 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_190333-82mnef5m/files/output.log +2024-06-08 19:19:14,000 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:19:14,273 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 19:19:15,000 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:19:16,000 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:19:17,000 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:19:18,000 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:19:19,000 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:19:19,364 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: stop_status +2024-06-08 19:19:19,364 DEBUG SenderThread:30410 [sender.py:send_request():406] send_request: stop_status +2024-06-08 19:19:19,430 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 19:19:20,000 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:19:21,000 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:19:22,000 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:19:23,000 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:19:24,000 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:19:24,219 INFO Thread-12 :30410 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_190333-82mnef5m/files/output.log +2024-06-08 19:19:25,001 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:19:25,134 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 19:19:26,001 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:19:26,403 INFO Thread-12 :30410 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_190333-82mnef5m/files/output.log +2024-06-08 19:19:27,001 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:19:28,001 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:19:29,001 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:19:30,001 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:19:30,789 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 19:19:31,001 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:19:32,001 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:19:33,001 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:19:34,001 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:19:34,146 DEBUG SenderThread:30410 [sender.py:send():379] send: stats +2024-06-08 19:19:34,364 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: stop_status +2024-06-08 19:19:34,364 DEBUG SenderThread:30410 [sender.py:send_request():406] send_request: stop_status +2024-06-08 19:19:35,001 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:19:35,789 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 19:19:36,002 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:19:37,001 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:19:38,002 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:19:38,051 INFO Thread-12 :30410 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_190333-82mnef5m/files/output.log +2024-06-08 19:19:39,002 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:19:40,002 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:19:41,002 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:19:41,789 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 19:19:42,002 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:19:43,002 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:19:44,002 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:19:45,002 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:19:46,002 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:19:46,789 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 19:19:47,002 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:19:48,003 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:19:49,003 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:19:49,364 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: stop_status +2024-06-08 19:19:49,364 DEBUG SenderThread:30410 [sender.py:send_request():406] send_request: stop_status +2024-06-08 19:19:49,751 INFO Thread-12 :30410 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_190333-82mnef5m/files/output.log +2024-06-08 19:19:50,003 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:19:51,003 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:19:51,790 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 19:19:51,926 INFO Thread-12 :30410 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_190333-82mnef5m/files/output.log +2024-06-08 19:19:52,003 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:19:53,003 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:19:54,003 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:19:55,003 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:19:56,003 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:19:56,790 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 19:19:57,003 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:19:58,003 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:19:59,004 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:20:00,004 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:20:01,004 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:20:01,850 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 19:20:02,004 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:20:03,004 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:20:03,753 INFO Thread-12 :30410 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_190333-82mnef5m/files/output.log +2024-06-08 19:20:04,004 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:20:04,148 DEBUG SenderThread:30410 [sender.py:send():379] send: stats +2024-06-08 19:20:04,364 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: stop_status +2024-06-08 19:20:04,364 DEBUG SenderThread:30410 [sender.py:send_request():406] send_request: stop_status +2024-06-08 19:20:05,004 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:20:06,004 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:20:07,004 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:20:07,790 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 19:20:08,004 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:20:09,004 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:20:10,005 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:20:11,005 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:20:12,005 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:20:13,005 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:20:13,284 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 19:20:14,005 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:20:14,265 INFO Thread-12 :30410 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_190333-82mnef5m/files/output.log +2024-06-08 19:20:15,005 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:20:16,005 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:20:16,519 INFO Thread-12 :30410 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_190333-82mnef5m/files/output.log +2024-06-08 19:20:17,005 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:20:17,771 INFO Thread-12 :30410 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_190333-82mnef5m/files/output.log +2024-06-08 19:20:18,005 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:20:18,790 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 19:20:19,005 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:20:19,364 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: stop_status +2024-06-08 19:20:19,364 DEBUG SenderThread:30410 [sender.py:send_request():406] send_request: stop_status +2024-06-08 19:20:20,005 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:20:21,006 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:20:22,006 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:20:23,006 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:20:23,790 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 19:20:24,006 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:20:25,006 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:20:26,006 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:20:27,006 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:20:28,006 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:20:29,006 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:20:29,536 INFO Thread-12 :30410 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_190333-82mnef5m/files/output.log +2024-06-08 19:20:29,626 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 19:20:30,006 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:20:31,006 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:20:32,007 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:20:33,007 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:20:34,007 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:20:34,150 DEBUG SenderThread:30410 [sender.py:send():379] send: stats +2024-06-08 19:20:34,364 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: stop_status +2024-06-08 19:20:34,365 DEBUG SenderThread:30410 [sender.py:send_request():406] send_request: stop_status +2024-06-08 19:20:35,007 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:20:35,504 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 19:20:36,007 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:20:37,007 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:20:38,007 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:20:39,007 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:20:40,007 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:20:40,175 INFO Thread-12 :30410 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_190333-82mnef5m/files/output.log +2024-06-08 19:20:40,602 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 19:20:41,007 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:20:42,007 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:20:42,313 INFO Thread-12 :30410 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_190333-82mnef5m/files/output.log +2024-06-08 19:20:43,008 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:20:44,008 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:20:45,008 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:20:45,791 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 19:20:46,008 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:20:47,008 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:20:48,008 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:20:49,008 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:20:49,364 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: stop_status +2024-06-08 19:20:49,365 DEBUG SenderThread:30410 [sender.py:send_request():406] send_request: stop_status +2024-06-08 19:20:50,008 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:20:50,791 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 19:20:51,008 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:20:52,008 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:20:53,008 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:20:54,008 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:20:54,299 INFO Thread-12 :30410 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_190333-82mnef5m/files/output.log +2024-06-08 19:20:55,009 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:20:55,583 INFO Thread-12 :30410 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_190333-82mnef5m/files/output.log +2024-06-08 19:20:55,791 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 19:20:56,009 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:20:57,009 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:20:58,009 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:20:59,009 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:21:00,009 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:21:00,791 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 19:21:01,009 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:21:02,009 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:21:03,009 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:21:04,009 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:21:04,152 DEBUG SenderThread:30410 [sender.py:send():379] send: stats +2024-06-08 19:21:04,364 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: stop_status +2024-06-08 19:21:04,365 DEBUG SenderThread:30410 [sender.py:send_request():406] send_request: stop_status +2024-06-08 19:21:05,009 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:21:05,791 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 19:21:06,010 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:21:07,010 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:21:08,010 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:21:08,135 INFO Thread-12 :30410 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_190333-82mnef5m/files/output.log +2024-06-08 19:21:09,010 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:21:10,010 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:21:10,791 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 19:21:11,010 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:21:12,010 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:21:13,010 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:21:14,010 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:21:15,010 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:21:15,791 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 19:21:16,010 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:21:17,010 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:21:18,011 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:21:19,011 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:21:19,364 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: stop_status +2024-06-08 19:21:19,365 DEBUG SenderThread:30410 [sender.py:send_request():406] send_request: stop_status +2024-06-08 19:21:20,011 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:21:20,679 INFO Thread-12 :30410 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_190333-82mnef5m/files/output.log +2024-06-08 19:21:20,791 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 19:21:21,011 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:21:21,875 INFO Thread-12 :30410 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_190333-82mnef5m/files/output.log +2024-06-08 19:21:22,011 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:21:23,011 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:21:24,011 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:21:25,011 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:21:25,792 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 19:21:26,011 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:21:27,011 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:21:28,011 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:21:29,012 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:21:30,012 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:21:30,792 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 19:21:31,012 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:21:32,012 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:21:33,012 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:21:33,915 INFO Thread-12 :30410 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_190333-82mnef5m/files/output.log +2024-06-08 19:21:34,012 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:21:34,168 DEBUG SenderThread:30410 [sender.py:send():379] send: stats +2024-06-08 19:21:34,365 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: stop_status +2024-06-08 19:21:34,365 DEBUG SenderThread:30410 [sender.py:send_request():406] send_request: stop_status +2024-06-08 19:21:35,012 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:21:36,012 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:21:36,792 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 19:21:37,012 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:21:38,013 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:21:39,013 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:21:40,013 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:21:41,013 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:21:42,013 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:21:42,792 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 19:21:43,013 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:21:44,013 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:21:45,013 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:21:46,013 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:21:46,034 INFO Thread-12 :30410 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_190333-82mnef5m/files/output.log +2024-06-08 19:21:47,013 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:21:48,013 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:21:48,792 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 19:21:49,013 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:21:49,365 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: stop_status +2024-06-08 19:21:49,365 DEBUG SenderThread:30410 [sender.py:send_request():406] send_request: stop_status +2024-06-08 19:21:50,014 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:21:51,014 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:21:52,014 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:21:53,014 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:21:53,792 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 19:21:54,014 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:21:55,014 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:21:56,014 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:21:57,014 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:21:58,014 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:21:58,892 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 19:21:59,014 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:22:00,014 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:22:00,116 INFO Thread-12 :30410 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_190333-82mnef5m/files/output.log +2024-06-08 19:22:01,014 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:22:02,015 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:22:03,015 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:22:04,015 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:22:04,171 DEBUG SenderThread:30410 [sender.py:send():379] send: stats +2024-06-08 19:22:04,173 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 19:22:04,365 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: stop_status +2024-06-08 19:22:04,365 DEBUG SenderThread:30410 [sender.py:send_request():406] send_request: stop_status +2024-06-08 19:22:05,015 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:22:06,015 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:22:07,015 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:22:08,015 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:22:09,015 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:22:09,183 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 19:22:10,015 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:22:10,659 INFO Thread-12 :30410 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_190333-82mnef5m/files/output.log +2024-06-08 19:22:11,015 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:22:11,787 INFO Thread-12 :30410 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_190333-82mnef5m/files/output.log +2024-06-08 19:22:12,016 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:22:13,016 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:22:14,016 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:22:15,016 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:22:15,180 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 19:22:16,016 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:22:17,016 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:22:18,016 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:22:19,016 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:22:19,365 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: stop_status +2024-06-08 19:22:19,365 DEBUG SenderThread:30410 [sender.py:send_request():406] send_request: stop_status +2024-06-08 19:22:20,016 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:22:20,534 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 19:22:21,016 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:22:22,016 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:22:23,016 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:22:24,017 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:22:24,659 INFO Thread-12 :30410 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_190333-82mnef5m/files/output.log +2024-06-08 19:22:25,017 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:22:25,793 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 19:22:26,017 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:22:27,017 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:22:28,017 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:22:29,017 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:22:30,017 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:22:31,017 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:22:31,793 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 19:22:32,017 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:22:33,017 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:22:34,018 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:22:34,172 DEBUG SenderThread:30410 [sender.py:send():379] send: stats +2024-06-08 19:22:34,365 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: stop_status +2024-06-08 19:22:34,365 DEBUG SenderThread:30410 [sender.py:send_request():406] send_request: stop_status +2024-06-08 19:22:35,018 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:22:36,025 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:22:37,025 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:22:37,489 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 19:22:38,025 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:22:38,603 INFO Thread-12 :30410 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_190333-82mnef5m/files/output.log +2024-06-08 19:22:39,025 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:22:40,025 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:22:41,025 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:22:42,026 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:22:42,793 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 19:22:43,026 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:22:44,026 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:22:45,026 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:22:46,026 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:22:47,026 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:22:48,026 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:22:48,793 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 19:22:49,026 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:22:49,365 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: stop_status +2024-06-08 19:22:49,365 DEBUG SenderThread:30410 [sender.py:send_request():406] send_request: stop_status +2024-06-08 19:22:50,026 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:22:50,323 INFO Thread-12 :30410 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_190333-82mnef5m/files/output.log +2024-06-08 19:22:51,026 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:22:52,027 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:22:52,335 INFO Thread-12 :30410 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_190333-82mnef5m/files/output.log +2024-06-08 19:22:53,027 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:22:54,027 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:22:54,793 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 19:22:55,027 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:22:56,027 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:22:57,027 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:22:58,027 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:22:59,027 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:22:59,793 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 19:23:00,027 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:23:01,027 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:23:02,027 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:23:02,556 INFO Thread-12 :30410 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_190333-82mnef5m/files/output.log +2024-06-08 19:23:03,028 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:23:04,028 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:23:04,174 DEBUG SenderThread:30410 [sender.py:send():379] send: stats +2024-06-08 19:23:04,365 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: stop_status +2024-06-08 19:23:04,365 DEBUG SenderThread:30410 [sender.py:send_request():406] send_request: stop_status +2024-06-08 19:23:04,883 INFO Thread-12 :30410 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_190333-82mnef5m/files/output.log +2024-06-08 19:23:05,028 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:23:05,471 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 19:23:06,028 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:23:07,028 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:23:08,028 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:23:09,028 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:23:10,028 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:23:10,793 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 19:23:11,028 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:23:12,028 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:23:13,029 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:23:14,003 INFO Thread-12 :30410 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_190333-82mnef5m/files/output.log +2024-06-08 19:23:14,029 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:23:15,029 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:23:15,793 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 19:23:16,029 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:23:16,355 INFO Thread-12 :30410 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_190333-82mnef5m/files/output.log +2024-06-08 19:23:17,029 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:23:18,029 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:23:19,029 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:23:19,365 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: stop_status +2024-06-08 19:23:19,365 DEBUG SenderThread:30410 [sender.py:send_request():406] send_request: stop_status +2024-06-08 19:23:20,029 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:23:20,794 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 19:23:21,029 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:23:22,029 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:23:23,029 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:23:24,030 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:23:25,030 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:23:25,794 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 19:23:26,030 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:23:27,030 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:23:28,030 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:23:29,030 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:23:30,030 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:23:30,743 INFO Thread-12 :30410 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_190333-82mnef5m/files/output.log +2024-06-08 19:23:31,030 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:23:31,794 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 19:23:32,030 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:23:33,030 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:23:34,030 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:23:34,177 DEBUG SenderThread:30410 [sender.py:send():379] send: stats +2024-06-08 19:23:34,365 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: stop_status +2024-06-08 19:23:34,366 DEBUG SenderThread:30410 [sender.py:send_request():406] send_request: stop_status +2024-06-08 19:23:35,031 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:23:36,030 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:23:36,794 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 19:23:37,031 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:23:38,031 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:23:39,031 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:23:40,031 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:23:40,037 INFO Thread-12 :30410 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_190333-82mnef5m/files/output.log +2024-06-08 19:23:41,031 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:23:42,031 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:23:42,794 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 19:23:43,031 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:23:44,031 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:23:45,031 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:23:46,031 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:23:47,032 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:23:47,795 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 19:23:48,032 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:23:49,032 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:23:49,365 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: stop_status +2024-06-08 19:23:49,366 DEBUG SenderThread:30410 [sender.py:send_request():406] send_request: stop_status +2024-06-08 19:23:50,032 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:23:50,239 INFO Thread-12 :30410 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_190333-82mnef5m/files/output.log +2024-06-08 19:23:51,032 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:23:52,032 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:23:53,032 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:23:53,794 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 19:23:54,032 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:23:55,032 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:23:56,032 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:23:57,033 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:23:58,033 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:23:59,033 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:23:59,588 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 19:23:59,863 INFO Thread-12 :30410 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_190333-82mnef5m/files/output.log +2024-06-08 19:24:00,038 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:24:01,038 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:24:02,038 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:24:03,038 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 19:24:04,179 DEBUG SenderThread:30410 [sender.py:send():379] send: stats +2024-06-08 19:24:04,366 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: stop_status +2024-06-08 19:24:04,598 DEBUG SenderThread:30410 [sender.py:send_request():406] send_request: stop_status +2024-06-08 19:24:04,758 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 19:24:09,923 INFO Thread-12 :30410 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_190333-82mnef5m/files/output.log +2024-06-08 19:24:10,350 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 19:24:15,795 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 19:24:19,365 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: stop_status +2024-06-08 19:24:19,366 DEBUG SenderThread:30410 [sender.py:send_request():406] send_request: stop_status +2024-06-08 19:24:20,307 INFO Thread-12 :30410 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_190333-82mnef5m/files/output.log +2024-06-08 19:24:21,795 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 19:24:26,795 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 19:24:30,464 INFO Thread-12 :30410 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_190333-82mnef5m/files/output.log +2024-06-08 19:24:32,627 INFO Thread-12 :30410 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_190333-82mnef5m/files/output.log +2024-06-08 19:24:32,795 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 19:24:34,181 DEBUG SenderThread:30410 [sender.py:send():379] send: stats +2024-06-08 19:24:34,366 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: stop_status +2024-06-08 19:24:34,366 DEBUG SenderThread:30410 [sender.py:send_request():406] send_request: stop_status +2024-06-08 19:24:37,795 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 19:24:42,871 INFO Thread-12 :30410 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_190333-82mnef5m/files/output.log +2024-06-08 19:24:43,795 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 19:24:48,795 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 19:24:49,366 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: stop_status +2024-06-08 19:24:49,366 DEBUG SenderThread:30410 [sender.py:send_request():406] send_request: stop_status +2024-06-08 19:24:52,647 INFO Thread-12 :30410 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_190333-82mnef5m/files/output.log +2024-06-08 19:24:54,795 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 19:24:59,795 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 19:25:01,875 INFO Thread-12 :30410 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_190333-82mnef5m/files/output.log +2024-06-08 19:25:04,183 DEBUG SenderThread:30410 [sender.py:send():379] send: stats +2024-06-08 19:25:04,366 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: stop_status +2024-06-08 19:25:04,366 DEBUG SenderThread:30410 [sender.py:send_request():406] send_request: stop_status +2024-06-08 19:25:05,530 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 19:25:10,561 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 19:25:12,523 INFO Thread-12 :30410 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_190333-82mnef5m/files/output.log +2024-06-08 19:25:15,796 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 19:25:19,366 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: stop_status +2024-06-08 19:25:19,366 DEBUG SenderThread:30410 [sender.py:send_request():406] send_request: stop_status +2024-06-08 19:25:20,815 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 19:25:22,951 INFO Thread-12 :30410 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_190333-82mnef5m/files/output.log +2024-06-08 19:25:26,796 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 19:25:31,976 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 19:25:32,115 INFO Thread-12 :30410 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_190333-82mnef5m/files/output.log +2024-06-08 19:25:34,185 DEBUG SenderThread:30410 [sender.py:send():379] send: stats +2024-06-08 19:25:34,366 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: stop_status +2024-06-08 19:25:34,366 DEBUG SenderThread:30410 [sender.py:send_request():406] send_request: stop_status +2024-06-08 19:25:37,796 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 19:25:42,650 INFO Thread-12 :30410 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_190333-82mnef5m/files/output.log +2024-06-08 19:25:43,108 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 19:25:48,796 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 19:25:49,366 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: stop_status +2024-06-08 19:25:49,366 DEBUG SenderThread:30410 [sender.py:send_request():406] send_request: stop_status +2024-06-08 19:25:52,892 INFO Thread-12 :30410 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_190333-82mnef5m/files/output.log +2024-06-08 19:25:54,213 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 19:25:59,806 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 19:26:02,635 INFO Thread-12 :30410 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_190333-82mnef5m/files/output.log +2024-06-08 19:26:04,187 DEBUG SenderThread:30410 [sender.py:send():379] send: stats +2024-06-08 19:26:04,366 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: stop_status +2024-06-08 19:26:04,367 DEBUG SenderThread:30410 [sender.py:send_request():406] send_request: stop_status +2024-06-08 19:26:05,465 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 19:26:10,806 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 19:26:12,707 INFO Thread-12 :30410 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_190333-82mnef5m/files/output.log +2024-06-08 19:26:15,806 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 19:26:19,366 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: stop_status +2024-06-08 19:26:19,367 DEBUG SenderThread:30410 [sender.py:send_request():406] send_request: stop_status +2024-06-08 19:26:20,806 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 19:26:22,147 INFO Thread-12 :30410 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_190333-82mnef5m/files/output.log +2024-06-08 19:26:25,806 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 19:26:30,807 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 19:26:32,756 INFO Thread-12 :30410 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_190333-82mnef5m/files/output.log +2024-06-08 19:26:34,189 DEBUG SenderThread:30410 [sender.py:send():379] send: stats +2024-06-08 19:26:34,366 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: stop_status +2024-06-08 19:26:34,367 DEBUG SenderThread:30410 [sender.py:send_request():406] send_request: stop_status +2024-06-08 19:26:36,807 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 19:26:42,335 INFO Thread-12 :30410 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_190333-82mnef5m/files/output.log +2024-06-08 19:26:42,494 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 19:26:47,807 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 19:26:49,366 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: stop_status +2024-06-08 19:26:49,367 DEBUG SenderThread:30410 [sender.py:send_request():406] send_request: stop_status +2024-06-08 19:26:52,691 INFO Thread-12 :30410 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_190333-82mnef5m/files/output.log +2024-06-08 19:26:53,742 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 19:26:55,065 INFO Thread-12 :30410 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_190333-82mnef5m/files/output.log +2024-06-08 19:26:58,807 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 19:27:02,266 INFO Thread-12 :30410 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_190333-82mnef5m/files/output.log +2024-06-08 19:27:04,191 DEBUG SenderThread:30410 [sender.py:send():379] send: stats +2024-06-08 19:27:04,192 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 19:27:04,367 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: stop_status +2024-06-08 19:27:04,367 DEBUG SenderThread:30410 [sender.py:send_request():406] send_request: stop_status +2024-06-08 19:27:09,807 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 19:27:13,048 INFO Thread-12 :30410 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_190333-82mnef5m/files/output.log +2024-06-08 19:27:15,691 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 19:27:19,367 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: stop_status +2024-06-08 19:27:19,367 DEBUG SenderThread:30410 [sender.py:send_request():406] send_request: stop_status +2024-06-08 19:27:20,807 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 19:27:22,739 INFO Thread-12 :30410 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_190333-82mnef5m/files/output.log +2024-06-08 19:27:25,024 INFO Thread-12 :30410 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_190333-82mnef5m/files/output.log +2024-06-08 19:27:26,807 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 19:27:31,894 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 19:27:33,297 INFO Thread-12 :30410 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_190333-82mnef5m/files/output.log +2024-06-08 19:27:34,193 DEBUG SenderThread:30410 [sender.py:send():379] send: stats +2024-06-08 19:27:34,367 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: stop_status +2024-06-08 19:27:34,367 DEBUG SenderThread:30410 [sender.py:send_request():406] send_request: stop_status +2024-06-08 19:27:34,564 INFO Thread-12 :30410 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_190333-82mnef5m/files/output.log +2024-06-08 19:27:37,808 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 19:27:42,862 INFO Thread-12 :30410 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_190333-82mnef5m/files/output.log +2024-06-08 19:27:42,997 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 19:27:45,076 INFO Thread-12 :30410 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_190333-82mnef5m/files/output.log +2024-06-08 19:27:48,808 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 19:27:49,367 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: stop_status +2024-06-08 19:27:49,367 DEBUG SenderThread:30410 [sender.py:send_request():406] send_request: stop_status +2024-06-08 19:27:52,956 INFO Thread-12 :30410 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_190333-82mnef5m/files/output.log +2024-06-08 19:27:54,165 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 19:27:55,168 INFO Thread-12 :30410 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_190333-82mnef5m/files/output.log +2024-06-08 19:27:59,808 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 19:28:02,118 INFO Thread-12 :30410 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_190333-82mnef5m/files/output.log +2024-06-08 19:28:04,195 DEBUG SenderThread:30410 [sender.py:send():379] send: stats +2024-06-08 19:28:04,367 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: stop_status +2024-06-08 19:28:04,367 DEBUG SenderThread:30410 [sender.py:send_request():406] send_request: stop_status +2024-06-08 19:28:04,379 INFO Thread-12 :30410 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_190333-82mnef5m/files/output.log +2024-06-08 19:28:05,498 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 19:28:10,808 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 19:28:14,671 INFO Thread-12 :30410 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_190333-82mnef5m/files/output.log +2024-06-08 19:28:15,808 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 19:28:19,367 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: stop_status +2024-06-08 19:28:19,367 DEBUG SenderThread:30410 [sender.py:send_request():406] send_request: stop_status +2024-06-08 19:28:20,809 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 19:28:22,694 INFO Thread-12 :30410 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_190333-82mnef5m/files/output.log +2024-06-08 19:28:24,996 INFO Thread-12 :30410 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_190333-82mnef5m/files/output.log +2024-06-08 19:28:25,809 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 19:28:31,668 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 19:28:32,371 INFO Thread-12 :30410 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_190333-82mnef5m/files/output.log +2024-06-08 19:28:34,197 DEBUG SenderThread:30410 [sender.py:send():379] send: stats +2024-06-08 19:28:34,367 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: stop_status +2024-06-08 19:28:34,367 DEBUG SenderThread:30410 [sender.py:send_request():406] send_request: stop_status +2024-06-08 19:28:34,747 INFO Thread-12 :30410 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_190333-82mnef5m/files/output.log +2024-06-08 19:28:36,809 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 19:28:41,897 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 19:28:43,048 INFO Thread-12 :30410 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_190333-82mnef5m/files/output.log +2024-06-08 19:28:45,260 INFO Thread-12 :30410 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_190333-82mnef5m/files/output.log +2024-06-08 19:28:47,809 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 19:28:49,367 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: stop_status +2024-06-08 19:28:49,367 DEBUG SenderThread:30410 [sender.py:send_request():406] send_request: stop_status +2024-06-08 19:28:52,323 INFO Thread-12 :30410 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_190333-82mnef5m/files/output.log +2024-06-08 19:28:53,424 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 19:28:58,809 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 19:29:02,435 INFO Thread-12 :30410 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_190333-82mnef5m/files/output.log +2024-06-08 19:29:03,908 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 19:29:04,199 DEBUG SenderThread:30410 [sender.py:send():379] send: stats +2024-06-08 19:29:04,367 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: stop_status +2024-06-08 19:29:04,368 DEBUG SenderThread:30410 [sender.py:send_request():406] send_request: stop_status +2024-06-08 19:29:09,810 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 19:29:12,815 INFO Thread-12 :30410 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_190333-82mnef5m/files/output.log +2024-06-08 19:29:15,810 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 19:29:19,367 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: stop_status +2024-06-08 19:29:19,367 DEBUG SenderThread:30410 [sender.py:send_request():406] send_request: stop_status +2024-06-08 19:29:20,487 INFO Thread-12 :30410 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_190333-82mnef5m/files/output.log +2024-06-08 19:29:20,810 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 19:29:25,810 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 19:29:30,571 INFO Thread-12 :30410 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_190333-82mnef5m/files/output.log +2024-06-08 19:29:31,810 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 19:29:34,201 DEBUG SenderThread:30410 [sender.py:send():379] send: stats +2024-06-08 19:29:34,367 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: stop_status +2024-06-08 19:29:34,368 DEBUG SenderThread:30410 [sender.py:send_request():406] send_request: stop_status +2024-06-08 19:29:37,293 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 19:29:39,010 INFO Thread-12 :30410 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_190333-82mnef5m/files/output.log +2024-06-08 19:29:42,810 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 19:29:47,857 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 19:29:48,227 INFO Thread-12 :30410 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_190333-82mnef5m/files/output.log +2024-06-08 19:29:49,368 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: stop_status +2024-06-08 19:29:49,368 DEBUG SenderThread:30410 [sender.py:send_request():406] send_request: stop_status +2024-06-08 19:29:53,810 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 19:29:56,571 INFO Thread-12 :30410 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_190333-82mnef5m/files/output.log +2024-06-08 19:29:58,923 INFO Thread-12 :30410 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_190333-82mnef5m/files/output.log +2024-06-08 19:29:59,067 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 19:30:04,203 DEBUG SenderThread:30410 [sender.py:send():379] send: stats +2024-06-08 19:30:04,204 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 19:30:04,368 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: stop_status +2024-06-08 19:30:04,368 DEBUG SenderThread:30410 [sender.py:send_request():406] send_request: stop_status +2024-06-08 19:30:06,842 INFO Thread-12 :30410 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_190333-82mnef5m/files/output.log +2024-06-08 19:30:09,815 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 19:30:15,427 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 19:30:17,092 INFO Thread-12 :30410 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_190333-82mnef5m/files/output.log +2024-06-08 19:30:19,368 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: stop_status +2024-06-08 19:30:19,368 DEBUG SenderThread:30410 [sender.py:send_request():406] send_request: stop_status +2024-06-08 19:30:20,471 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 19:30:25,129 INFO Thread-12 :30410 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_190333-82mnef5m/files/output.log +2024-06-08 19:30:25,581 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 19:30:30,811 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 19:30:34,206 DEBUG SenderThread:30410 [sender.py:send():379] send: stats +2024-06-08 19:30:34,368 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: stop_status +2024-06-08 19:30:34,368 DEBUG SenderThread:30410 [sender.py:send_request():406] send_request: stop_status +2024-06-08 19:30:35,256 INFO Thread-12 :30410 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_190333-82mnef5m/files/output.log +2024-06-08 19:30:35,811 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 19:30:40,811 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 19:30:43,162 INFO Thread-12 :30410 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_190333-82mnef5m/files/output.log +2024-06-08 19:30:44,307 INFO Thread-12 :30410 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_190333-82mnef5m/files/output.log +2024-06-08 19:30:45,811 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 19:30:49,368 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: stop_status +2024-06-08 19:30:49,368 DEBUG SenderThread:30410 [sender.py:send_request():406] send_request: stop_status +2024-06-08 19:30:50,811 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 19:30:52,703 INFO Thread-12 :30410 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_190333-82mnef5m/files/output.log +2024-06-08 19:30:56,811 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 19:31:02,811 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 19:31:03,296 INFO Thread-12 :30410 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_190333-82mnef5m/files/output.log +2024-06-08 19:31:04,208 DEBUG SenderThread:30410 [sender.py:send():379] send: stats +2024-06-08 19:31:04,368 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: stop_status +2024-06-08 19:31:04,368 DEBUG SenderThread:30410 [sender.py:send_request():406] send_request: stop_status +2024-06-08 19:31:07,812 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 19:31:11,335 INFO Thread-12 :30410 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_190333-82mnef5m/files/output.log +2024-06-08 19:31:13,812 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 19:31:18,907 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 19:31:19,368 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: stop_status +2024-06-08 19:31:19,368 DEBUG SenderThread:30410 [sender.py:send_request():406] send_request: stop_status +2024-06-08 19:31:20,741 INFO Thread-12 :30410 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_190333-82mnef5m/files/output.log +2024-06-08 19:31:24,812 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 19:31:30,452 INFO Thread-12 :30410 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_190333-82mnef5m/files/output.log +2024-06-08 19:31:30,587 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 19:31:34,210 DEBUG SenderThread:30410 [sender.py:send():379] send: stats +2024-06-08 19:31:34,368 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: stop_status +2024-06-08 19:31:34,368 DEBUG SenderThread:30410 [sender.py:send_request():406] send_request: stop_status +2024-06-08 19:31:35,812 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 19:31:38,899 INFO Thread-12 :30410 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_190333-82mnef5m/files/output.log +2024-06-08 19:31:41,812 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 19:31:46,812 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 19:31:49,368 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: stop_status +2024-06-08 19:31:49,369 DEBUG SenderThread:30410 [sender.py:send_request():406] send_request: stop_status +2024-06-08 19:31:49,464 INFO Thread-12 :30410 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_190333-82mnef5m/files/output.log +2024-06-08 19:31:52,812 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 19:31:58,635 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 19:31:58,768 INFO Thread-12 :30410 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_190333-82mnef5m/files/output.log +2024-06-08 19:32:03,813 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 19:32:04,212 DEBUG SenderThread:30410 [sender.py:send():379] send: stats +2024-06-08 19:32:04,368 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: stop_status +2024-06-08 19:32:04,369 DEBUG SenderThread:30410 [sender.py:send_request():406] send_request: stop_status +2024-06-08 19:32:06,679 INFO Thread-12 :30410 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_190333-82mnef5m/files/output.log +2024-06-08 19:32:09,813 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 19:32:14,890 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 19:32:17,269 INFO Thread-12 :30410 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_190333-82mnef5m/files/output.log +2024-06-08 19:32:19,368 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: stop_status +2024-06-08 19:32:19,369 DEBUG SenderThread:30410 [sender.py:send_request():406] send_request: stop_status +2024-06-08 19:32:20,483 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 19:32:24,335 INFO Thread-12 :30410 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_190333-82mnef5m/files/output.log +2024-06-08 19:32:25,813 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 19:32:26,940 INFO Thread-12 :30410 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_190333-82mnef5m/files/output.log +2024-06-08 19:32:30,813 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 19:32:34,214 DEBUG SenderThread:30410 [sender.py:send():379] send: stats +2024-06-08 19:32:34,368 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: stop_status +2024-06-08 19:32:34,369 DEBUG SenderThread:30410 [sender.py:send_request():406] send_request: stop_status +2024-06-08 19:32:35,218 INFO Thread-12 :30410 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_190333-82mnef5m/files/output.log +2024-06-08 19:32:35,813 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 19:32:40,813 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 19:32:45,100 INFO Thread-12 :30410 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_190333-82mnef5m/files/output.log +2024-06-08 19:32:45,813 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 19:32:49,369 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: stop_status +2024-06-08 19:32:49,651 DEBUG SenderThread:30410 [sender.py:send_request():406] send_request: stop_status +2024-06-08 19:32:50,814 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 19:32:53,072 INFO Thread-12 :30410 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_190333-82mnef5m/files/output.log +2024-06-08 19:32:54,383 INFO Thread-12 :30410 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_190333-82mnef5m/files/output.log +2024-06-08 19:32:55,814 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 19:33:00,814 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 19:33:02,759 INFO Thread-12 :30410 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_190333-82mnef5m/files/output.log +2024-06-08 19:33:04,217 DEBUG SenderThread:30410 [sender.py:send():379] send: stats +2024-06-08 19:33:04,369 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: stop_status +2024-06-08 19:33:04,369 DEBUG SenderThread:30410 [sender.py:send_request():406] send_request: stop_status +2024-06-08 19:33:05,814 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 19:33:11,814 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 19:33:12,493 INFO Thread-12 :30410 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_190333-82mnef5m/files/output.log +2024-06-08 19:33:16,814 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 19:33:19,369 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: stop_status +2024-06-08 19:33:19,369 DEBUG SenderThread:30410 [sender.py:send_request():406] send_request: stop_status +2024-06-08 19:33:20,992 INFO Thread-12 :30410 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_190333-82mnef5m/files/output.log +2024-06-08 19:33:22,814 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 19:33:23,142 INFO Thread-12 :30410 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_190333-82mnef5m/files/output.log +2024-06-08 19:33:27,814 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 19:33:31,540 INFO Thread-12 :30410 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_190333-82mnef5m/files/output.log +2024-06-08 19:33:33,814 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 19:33:34,219 DEBUG SenderThread:30410 [sender.py:send():379] send: stats +2024-06-08 19:33:34,369 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: stop_status +2024-06-08 19:33:34,369 DEBUG SenderThread:30410 [sender.py:send_request():406] send_request: stop_status +2024-06-08 19:33:38,841 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 19:33:40,771 INFO Thread-12 :30410 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_190333-82mnef5m/files/output.log +2024-06-08 19:33:44,821 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 19:33:48,779 INFO Thread-12 :30410 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_190333-82mnef5m/files/output.log +2024-06-08 19:33:49,369 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: stop_status +2024-06-08 19:33:49,369 DEBUG SenderThread:30410 [sender.py:send_request():406] send_request: stop_status +2024-06-08 19:33:50,540 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 19:33:50,932 INFO Thread-12 :30410 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_190333-82mnef5m/files/output.log +2024-06-08 19:33:55,821 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 19:33:58,794 INFO Thread-12 :30410 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_190333-82mnef5m/files/output.log +2024-06-08 19:34:01,821 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 19:34:04,221 DEBUG SenderThread:30410 [sender.py:send():379] send: stats +2024-06-08 19:34:04,369 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: stop_status +2024-06-08 19:34:04,369 DEBUG SenderThread:30410 [sender.py:send_request():406] send_request: stop_status +2024-06-08 19:34:06,890 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 19:34:07,205 INFO Thread-12 :30410 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_190333-82mnef5m/files/output.log +2024-06-08 19:34:12,821 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 19:34:16,780 INFO Thread-12 :30410 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_190333-82mnef5m/files/output.log +2024-06-08 19:34:18,821 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 19:34:19,369 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: stop_status +2024-06-08 19:34:19,370 DEBUG SenderThread:30410 [sender.py:send_request():406] send_request: stop_status +2024-06-08 19:34:23,838 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 19:34:25,084 INFO Thread-12 :30410 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_190333-82mnef5m/files/output.log +2024-06-08 19:34:29,821 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 19:34:34,223 DEBUG SenderThread:30410 [sender.py:send():379] send: stats +2024-06-08 19:34:34,369 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: stop_status +2024-06-08 19:34:34,369 DEBUG SenderThread:30410 [sender.py:send_request():406] send_request: stop_status +2024-06-08 19:34:34,460 INFO Thread-12 :30410 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_190333-82mnef5m/files/output.log +2024-06-08 19:34:35,524 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 19:34:40,821 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 19:34:44,883 INFO Thread-12 :30410 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_190333-82mnef5m/files/output.log +2024-06-08 19:34:45,822 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 19:34:49,369 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: stop_status +2024-06-08 19:34:49,370 DEBUG SenderThread:30410 [sender.py:send_request():406] send_request: stop_status +2024-06-08 19:34:50,822 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 19:34:53,427 INFO Thread-12 :30410 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_190333-82mnef5m/files/output.log +2024-06-08 19:34:55,822 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 19:35:00,822 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 19:35:03,041 INFO Thread-12 :30410 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_190333-82mnef5m/files/output.log +2024-06-08 19:35:04,225 DEBUG SenderThread:30410 [sender.py:send():379] send: stats +2024-06-08 19:35:04,369 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: stop_status +2024-06-08 19:35:04,370 DEBUG SenderThread:30410 [sender.py:send_request():406] send_request: stop_status +2024-06-08 19:35:05,822 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 19:35:10,872 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 19:35:12,492 INFO Thread-12 :30410 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_190333-82mnef5m/files/output.log +2024-06-08 19:35:16,822 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 19:35:19,369 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: stop_status +2024-06-08 19:35:19,370 DEBUG SenderThread:30410 [sender.py:send_request():406] send_request: stop_status +2024-06-08 19:35:20,976 INFO Thread-12 :30410 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_190333-82mnef5m/files/output.log +2024-06-08 19:35:22,822 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 19:35:27,822 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 19:35:28,823 INFO Thread-12 :30410 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_190333-82mnef5m/files/output.log +2024-06-08 19:35:30,905 INFO Thread-12 :30410 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_190333-82mnef5m/files/output.log +2024-06-08 19:35:32,823 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 19:35:34,227 DEBUG SenderThread:30410 [sender.py:send():379] send: stats +2024-06-08 19:35:34,370 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: stop_status +2024-06-08 19:35:34,370 DEBUG SenderThread:30410 [sender.py:send_request():406] send_request: stop_status +2024-06-08 19:35:38,575 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 19:35:39,735 INFO Thread-12 :30410 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_190333-82mnef5m/files/output.log +2024-06-08 19:35:43,823 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 19:35:47,552 INFO Thread-12 :30410 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_190333-82mnef5m/files/output.log +2024-06-08 19:35:48,799 INFO Thread-12 :30410 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_190333-82mnef5m/files/output.log +2024-06-08 19:35:49,370 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: stop_status +2024-06-08 19:35:49,370 DEBUG SenderThread:30410 [sender.py:send_request():406] send_request: stop_status +2024-06-08 19:35:49,479 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 19:35:54,827 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 19:35:56,795 INFO Thread-12 :30410 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_190333-82mnef5m/files/output.log +2024-06-08 19:35:59,902 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 19:36:04,229 DEBUG SenderThread:30410 [sender.py:send():379] send: stats +2024-06-08 19:36:04,370 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: stop_status +2024-06-08 19:36:04,370 DEBUG SenderThread:30410 [sender.py:send_request():406] send_request: stop_status +2024-06-08 19:36:05,826 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 19:36:07,112 INFO Thread-12 :30410 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_190333-82mnef5m/files/output.log +2024-06-08 19:36:11,826 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 19:36:15,360 INFO Thread-12 :30410 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_190333-82mnef5m/files/output.log +2024-06-08 19:36:16,826 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 19:36:19,370 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: stop_status +2024-06-08 19:36:19,370 DEBUG SenderThread:30410 [sender.py:send_request():406] send_request: stop_status +2024-06-08 19:36:22,686 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 19:36:25,779 INFO Thread-12 :30410 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_190333-82mnef5m/files/output.log +2024-06-08 19:36:27,826 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 19:36:33,181 INFO Thread-12 :30410 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_190333-82mnef5m/files/output.log +2024-06-08 19:36:33,217 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 19:36:34,231 DEBUG SenderThread:30410 [sender.py:send():379] send: stats +2024-06-08 19:36:34,370 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: stop_status +2024-06-08 19:36:34,370 DEBUG SenderThread:30410 [sender.py:send_request():406] send_request: stop_status +2024-06-08 19:36:38,827 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 19:36:42,719 INFO Thread-12 :30410 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_190333-82mnef5m/files/output.log +2024-06-08 19:36:44,557 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 19:36:49,370 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: stop_status +2024-06-08 19:36:49,370 DEBUG SenderThread:30410 [sender.py:send_request():406] send_request: stop_status +2024-06-08 19:36:49,876 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 19:36:51,132 INFO Thread-12 :30410 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_190333-82mnef5m/files/output.log +2024-06-08 19:36:55,827 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 19:37:00,690 INFO Thread-12 :30410 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_190333-82mnef5m/files/output.log +2024-06-08 19:37:00,827 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 19:37:04,233 DEBUG SenderThread:30410 [sender.py:send():379] send: stats +2024-06-08 19:37:04,370 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: stop_status +2024-06-08 19:37:04,370 DEBUG SenderThread:30410 [sender.py:send_request():406] send_request: stop_status +2024-06-08 19:37:05,827 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 19:37:08,972 INFO Thread-12 :30410 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_190333-82mnef5m/files/output.log +2024-06-08 19:37:11,827 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 19:37:17,659 INFO Thread-12 :30410 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_190333-82mnef5m/files/output.log +2024-06-08 19:37:17,712 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 19:37:19,032 INFO Thread-12 :30410 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_190333-82mnef5m/files/output.log +2024-06-08 19:37:19,370 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: stop_status +2024-06-08 19:37:19,371 DEBUG SenderThread:30410 [sender.py:send_request():406] send_request: stop_status +2024-06-08 19:37:22,827 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 19:37:27,332 INFO Thread-12 :30410 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_190333-82mnef5m/files/output.log +2024-06-08 19:37:28,827 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 19:37:33,827 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 19:37:34,236 DEBUG SenderThread:30410 [sender.py:send():379] send: stats +2024-06-08 19:37:34,370 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: stop_status +2024-06-08 19:37:34,371 DEBUG SenderThread:30410 [sender.py:send_request():406] send_request: stop_status +2024-06-08 19:37:37,775 INFO Thread-12 :30410 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_190333-82mnef5m/files/output.log +2024-06-08 19:37:39,828 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 19:37:44,980 INFO Thread-12 :30410 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_190333-82mnef5m/files/output.log +2024-06-08 19:37:45,796 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 19:37:49,370 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: stop_status +2024-06-08 19:37:49,371 DEBUG SenderThread:30410 [sender.py:send_request():406] send_request: stop_status +2024-06-08 19:37:50,828 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 19:37:53,204 INFO Thread-12 :30410 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_190333-82mnef5m/files/output.log +2024-06-08 19:37:55,406 INFO Thread-12 :30410 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_190333-82mnef5m/files/output.log +2024-06-08 19:37:55,828 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 19:38:00,828 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 19:38:02,855 INFO Thread-12 :30410 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_190333-82mnef5m/files/output.log +2024-06-08 19:38:04,238 DEBUG SenderThread:30410 [sender.py:send():379] send: stats +2024-06-08 19:38:04,370 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: stop_status +2024-06-08 19:38:04,371 DEBUG SenderThread:30410 [sender.py:send_request():406] send_request: stop_status +2024-06-08 19:38:05,828 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 19:38:10,834 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 19:38:11,256 INFO Thread-12 :30410 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_190333-82mnef5m/files/output.log +2024-06-08 19:38:13,429 INFO Thread-12 :30410 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_190333-82mnef5m/files/output.log +2024-06-08 19:38:16,828 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 19:38:19,380 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: stop_status +2024-06-08 19:38:19,381 DEBUG SenderThread:30410 [sender.py:send_request():406] send_request: stop_status +2024-06-08 19:38:20,683 INFO Thread-12 :30410 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_190333-82mnef5m/files/output.log +2024-06-08 19:38:22,828 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 19:38:27,828 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 19:38:29,173 INFO Thread-12 :30410 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_190333-82mnef5m/files/output.log +2024-06-08 19:38:31,543 INFO Thread-12 :30410 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_190333-82mnef5m/files/output.log +2024-06-08 19:38:33,829 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 19:38:34,240 DEBUG SenderThread:30410 [sender.py:send():379] send: stats +2024-06-08 19:38:34,380 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: stop_status +2024-06-08 19:38:34,381 DEBUG SenderThread:30410 [sender.py:send_request():406] send_request: stop_status +2024-06-08 19:38:37,273 INFO Thread-12 :30410 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_190333-82mnef5m/files/output.log +2024-06-08 19:38:38,970 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 19:38:43,062 INFO Thread-12 :30410 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_190333-82mnef5m/files/output.log +2024-06-08 19:38:44,849 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 19:38:45,492 INFO Thread-12 :30410 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_190333-82mnef5m/files/output.log +2024-06-08 19:38:49,381 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: stop_status +2024-06-08 19:38:49,381 DEBUG SenderThread:30410 [sender.py:send_request():406] send_request: stop_status +2024-06-08 19:38:49,926 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 19:38:51,504 INFO Thread-12 :30410 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_190333-82mnef5m/files/output.log +2024-06-08 19:38:55,829 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 19:38:58,927 INFO Thread-12 :30410 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_190333-82mnef5m/files/output.log +2024-06-08 19:39:00,829 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 19:39:04,242 DEBUG SenderThread:30410 [sender.py:send():379] send: stats +2024-06-08 19:39:04,381 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: stop_status +2024-06-08 19:39:04,381 DEBUG SenderThread:30410 [sender.py:send_request():406] send_request: stop_status +2024-06-08 19:39:06,829 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 19:39:07,603 INFO Thread-12 :30410 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_190333-82mnef5m/files/output.log +2024-06-08 19:39:11,829 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 19:39:15,423 INFO Thread-12 :30410 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_190333-82mnef5m/files/output.log +2024-06-08 19:39:17,829 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 19:39:19,381 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: stop_status +2024-06-08 19:39:19,381 DEBUG SenderThread:30410 [sender.py:send_request():406] send_request: stop_status +2024-06-08 19:39:21,683 INFO Thread-12 :30410 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_190333-82mnef5m/files/output.log +2024-06-08 19:39:22,911 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 19:39:23,955 INFO Thread-12 :30410 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_190333-82mnef5m/files/output.log +2024-06-08 19:39:28,201 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 19:39:28,851 INFO Thread-12 :30410 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_190333-82mnef5m/files/output.log +2024-06-08 19:39:33,830 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 19:39:34,244 DEBUG SenderThread:30410 [sender.py:send():379] send: stats +2024-06-08 19:39:34,381 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: stop_status +2024-06-08 19:39:34,381 DEBUG SenderThread:30410 [sender.py:send_request():406] send_request: stop_status +2024-06-08 19:39:37,459 INFO Thread-12 :30410 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_190333-82mnef5m/files/output.log +2024-06-08 19:39:39,400 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 19:39:45,203 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 19:39:45,713 INFO Thread-12 :30410 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_190333-82mnef5m/files/output.log +2024-06-08 19:39:49,381 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: stop_status +2024-06-08 19:39:49,381 DEBUG SenderThread:30410 [sender.py:send_request():406] send_request: stop_status +2024-06-08 19:39:50,446 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 19:39:53,193 INFO Thread-12 :30410 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_190333-82mnef5m/files/output.log +2024-06-08 19:39:55,794 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 19:40:00,823 INFO Thread-12 :30410 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_190333-82mnef5m/files/output.log +2024-06-08 19:40:00,830 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 19:40:04,246 DEBUG SenderThread:30410 [sender.py:send():379] send: stats +2024-06-08 19:40:04,381 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: stop_status +2024-06-08 19:40:04,381 DEBUG SenderThread:30410 [sender.py:send_request():406] send_request: stop_status +2024-06-08 19:40:05,830 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 19:40:08,995 INFO Thread-12 :30410 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_190333-82mnef5m/files/output.log +2024-06-08 19:40:11,830 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 19:40:15,012 INFO Thread-12 :30410 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_190333-82mnef5m/files/output.log +2024-06-08 19:40:16,830 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 19:40:17,492 INFO Thread-12 :30410 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_190333-82mnef5m/files/output.log +2024-06-08 19:40:19,381 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: stop_status +2024-06-08 19:40:19,381 DEBUG SenderThread:30410 [sender.py:send_request():406] send_request: stop_status +2024-06-08 19:40:22,226 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 19:40:23,389 INFO Thread-12 :30410 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_190333-82mnef5m/files/output.log +2024-06-08 19:40:27,831 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 19:40:31,854 INFO Thread-12 :30410 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_190333-82mnef5m/files/output.log +2024-06-08 19:40:32,907 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 19:40:33,141 INFO Thread-12 :30410 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_190333-82mnef5m/files/output.log +2024-06-08 19:40:34,248 DEBUG SenderThread:30410 [sender.py:send():379] send: stats +2024-06-08 19:40:34,381 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: stop_status +2024-06-08 19:40:34,381 DEBUG SenderThread:30410 [sender.py:send_request():406] send_request: stop_status +2024-06-08 19:40:38,084 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 19:40:39,776 INFO Thread-12 :30410 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_190333-82mnef5m/files/output.log +2024-06-08 19:40:43,831 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 19:40:46,960 INFO Thread-12 :30410 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_190333-82mnef5m/files/output.log +2024-06-08 19:40:48,870 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 19:40:49,381 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: stop_status +2024-06-08 19:40:49,382 DEBUG SenderThread:30410 [sender.py:send_request():406] send_request: stop_status +2024-06-08 19:40:54,568 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 19:40:55,024 INFO Thread-12 :30410 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_190333-82mnef5m/files/output.log +2024-06-08 19:40:59,831 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 19:41:02,003 INFO Thread-12 :30410 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_190333-82mnef5m/files/output.log +2024-06-08 19:41:04,250 DEBUG SenderThread:30410 [sender.py:send():379] send: stats +2024-06-08 19:41:04,381 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: stop_status +2024-06-08 19:41:04,382 DEBUG SenderThread:30410 [sender.py:send_request():406] send_request: stop_status +2024-06-08 19:41:05,554 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 19:41:09,524 INFO Thread-12 :30410 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_190333-82mnef5m/files/output.log +2024-06-08 19:41:10,835 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 19:41:15,955 INFO Thread-12 :30410 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_190333-82mnef5m/files/output.log +2024-06-08 19:41:16,835 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 19:41:17,265 INFO Thread-12 :30410 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_190333-82mnef5m/files/output.log +2024-06-08 19:41:19,381 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: stop_status +2024-06-08 19:41:19,382 DEBUG SenderThread:30410 [sender.py:send_request():406] send_request: stop_status +2024-06-08 19:41:21,835 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 19:41:23,241 INFO Thread-12 :30410 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_190333-82mnef5m/files/output.log +2024-06-08 19:41:27,835 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 19:41:31,703 INFO Thread-12 :30410 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_190333-82mnef5m/files/output.log +2024-06-08 19:41:33,835 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 19:41:34,253 DEBUG SenderThread:30410 [sender.py:send():379] send: stats +2024-06-08 19:41:34,381 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: stop_status +2024-06-08 19:41:34,382 DEBUG SenderThread:30410 [sender.py:send_request():406] send_request: stop_status +2024-06-08 19:41:37,606 INFO Thread-12 :30410 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_190333-82mnef5m/files/output.log +2024-06-08 19:41:39,714 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 19:41:44,844 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 19:41:44,911 INFO Thread-12 :30410 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_190333-82mnef5m/files/output.log +2024-06-08 19:41:49,382 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: stop_status +2024-06-08 19:41:49,382 DEBUG SenderThread:30410 [sender.py:send_request():406] send_request: stop_status +2024-06-08 19:41:50,504 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 19:41:51,636 INFO Thread-12 :30410 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_190333-82mnef5m/files/output.log +2024-06-08 19:41:53,747 INFO Thread-12 :30410 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_190333-82mnef5m/files/output.log +2024-06-08 19:41:55,836 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 19:41:59,754 INFO Thread-12 :30410 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_190333-82mnef5m/files/output.log +2024-06-08 19:42:00,836 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 19:42:04,254 DEBUG SenderThread:30410 [sender.py:send():379] send: stats +2024-06-08 19:42:04,382 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: stop_status +2024-06-08 19:42:04,382 DEBUG SenderThread:30410 [sender.py:send_request():406] send_request: stop_status +2024-06-08 19:42:06,836 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 19:42:06,992 INFO Thread-12 :30410 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_190333-82mnef5m/files/output.log +2024-06-08 19:42:11,836 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 19:42:13,253 INFO Thread-12 :30410 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_190333-82mnef5m/files/output.log +2024-06-08 19:42:17,836 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 19:42:19,382 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: stop_status +2024-06-08 19:42:19,382 DEBUG SenderThread:30410 [sender.py:send_request():406] send_request: stop_status +2024-06-08 19:42:21,372 INFO Thread-12 :30410 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_190333-82mnef5m/files/output.log +2024-06-08 19:42:23,836 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 19:42:27,603 INFO Thread-12 :30410 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_190333-82mnef5m/files/output.log +2024-06-08 19:42:29,064 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 19:42:29,851 INFO Thread-12 :30410 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_190333-82mnef5m/files/output.log +2024-06-08 19:42:34,160 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 19:42:34,256 DEBUG SenderThread:30410 [sender.py:send():379] send: stats +2024-06-08 19:42:34,382 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: stop_status +2024-06-08 19:42:34,382 DEBUG SenderThread:30410 [sender.py:send_request():406] send_request: stop_status +2024-06-08 19:42:35,843 INFO Thread-12 :30410 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_190333-82mnef5m/files/output.log +2024-06-08 19:42:39,843 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 19:42:43,032 INFO Thread-12 :30410 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_190333-82mnef5m/files/output.log +2024-06-08 19:42:44,883 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 19:42:49,382 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: stop_status +2024-06-08 19:42:49,382 DEBUG SenderThread:30410 [sender.py:send_request():406] send_request: stop_status +2024-06-08 19:42:49,967 INFO Thread-12 :30410 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_190333-82mnef5m/files/output.log +2024-06-08 19:42:50,496 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 19:42:50,971 INFO Thread-12 :30410 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_190333-82mnef5m/files/output.log +2024-06-08 19:42:55,847 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 19:42:57,909 INFO Thread-12 :30410 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_190333-82mnef5m/files/output.log +2024-06-08 19:43:00,847 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 19:43:04,258 DEBUG SenderThread:30410 [sender.py:send():379] send: stats +2024-06-08 19:43:04,382 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: stop_status +2024-06-08 19:43:04,382 DEBUG SenderThread:30410 [sender.py:send_request():406] send_request: stop_status +2024-06-08 19:43:05,032 INFO Thread-12 :30410 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_190333-82mnef5m/files/output.log +2024-06-08 19:43:05,847 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 19:43:10,884 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 19:43:11,387 INFO Thread-12 :30410 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_190333-82mnef5m/files/output.log +2024-06-08 19:43:16,847 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 19:43:19,382 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: stop_status +2024-06-08 19:43:19,382 DEBUG SenderThread:30410 [sender.py:send_request():406] send_request: stop_status +2024-06-08 19:43:19,425 INFO Thread-12 :30410 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_190333-82mnef5m/files/output.log +2024-06-08 19:43:22,847 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 19:43:27,663 INFO Thread-12 :30410 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_190333-82mnef5m/files/output.log +2024-06-08 19:43:28,847 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 19:43:33,739 INFO Thread-12 :30410 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_190333-82mnef5m/files/output.log +2024-06-08 19:43:34,274 DEBUG SenderThread:30410 [sender.py:send():379] send: stats +2024-06-08 19:43:34,276 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 19:43:34,382 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: stop_status +2024-06-08 19:43:34,383 DEBUG SenderThread:30410 [sender.py:send_request():406] send_request: stop_status +2024-06-08 19:43:39,678 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 19:43:41,713 INFO Thread-12 :30410 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_190333-82mnef5m/files/output.log +2024-06-08 19:43:44,847 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 19:43:47,391 INFO Thread-12 :30410 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_190333-82mnef5m/files/output.log +2024-06-08 19:43:49,382 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: stop_status +2024-06-08 19:43:49,383 DEBUG SenderThread:30410 [sender.py:send_request():406] send_request: stop_status +2024-06-08 19:43:49,807 INFO Thread-12 :30410 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_190333-82mnef5m/files/output.log +2024-06-08 19:43:50,549 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 19:43:55,675 INFO Thread-12 :30410 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_190333-82mnef5m/files/output.log +2024-06-08 19:43:55,848 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 19:44:00,848 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 19:44:03,417 INFO Thread-12 :30410 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_190333-82mnef5m/files/output.log +2024-06-08 19:44:04,263 DEBUG SenderThread:30410 [sender.py:send():379] send: stats +2024-06-08 19:44:04,382 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: stop_status +2024-06-08 19:44:04,383 DEBUG SenderThread:30410 [sender.py:send_request():406] send_request: stop_status +2024-06-08 19:44:05,848 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 19:44:09,731 INFO Thread-12 :30410 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_190333-82mnef5m/files/output.log +2024-06-08 19:44:10,848 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 19:44:15,924 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 19:44:17,807 INFO Thread-12 :30410 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_190333-82mnef5m/files/output.log +2024-06-08 19:44:19,382 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: stop_status +2024-06-08 19:44:19,383 DEBUG SenderThread:30410 [sender.py:send_request():406] send_request: stop_status +2024-06-08 19:44:21,848 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 19:44:24,115 INFO Thread-12 :30410 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_190333-82mnef5m/files/output.log +2024-06-08 19:44:25,319 INFO Thread-12 :30410 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_190333-82mnef5m/files/output.log +2024-06-08 19:44:27,848 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 19:44:31,942 INFO Thread-12 :30410 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_190333-82mnef5m/files/output.log +2024-06-08 19:44:33,848 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 19:44:34,265 DEBUG SenderThread:30410 [sender.py:send():379] send: stats +2024-06-08 19:44:34,383 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: stop_status +2024-06-08 19:44:34,383 DEBUG SenderThread:30410 [sender.py:send_request():406] send_request: stop_status +2024-06-08 19:44:37,716 INFO Thread-12 :30410 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_190333-82mnef5m/files/output.log +2024-06-08 19:44:39,523 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 19:44:40,015 INFO Thread-12 :30410 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_190333-82mnef5m/files/output.log +2024-06-08 19:44:44,585 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 19:44:45,875 INFO Thread-12 :30410 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_190333-82mnef5m/files/output.log +2024-06-08 19:44:49,383 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: stop_status +2024-06-08 19:44:49,383 DEBUG SenderThread:30410 [sender.py:send_request():406] send_request: stop_status +2024-06-08 19:44:50,547 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 19:44:54,296 INFO Thread-12 :30410 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_190333-82mnef5m/files/output.log +2024-06-08 19:44:55,795 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 19:45:00,108 INFO Thread-12 :30410 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_190333-82mnef5m/files/output.log +2024-06-08 19:45:00,849 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 19:45:04,267 DEBUG SenderThread:30410 [sender.py:send():379] send: stats +2024-06-08 19:45:04,383 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: stop_status +2024-06-08 19:45:04,383 DEBUG SenderThread:30410 [sender.py:send_request():406] send_request: stop_status +2024-06-08 19:45:06,849 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 19:45:07,221 INFO Thread-12 :30410 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_190333-82mnef5m/files/output.log +2024-06-08 19:45:11,849 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 19:45:13,184 INFO Thread-12 :30410 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_190333-82mnef5m/files/output.log +2024-06-08 19:45:15,365 INFO Thread-12 :30410 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_190333-82mnef5m/files/output.log +2024-06-08 19:45:17,849 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 19:45:19,383 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: stop_status +2024-06-08 19:45:19,383 DEBUG SenderThread:30410 [sender.py:send_request():406] send_request: stop_status +2024-06-08 19:45:21,144 INFO Thread-12 :30410 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_190333-82mnef5m/files/output.log +2024-06-08 19:45:22,849 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 19:45:27,332 INFO Thread-12 :30410 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_190333-82mnef5m/files/output.log +2024-06-08 19:45:27,889 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 19:45:29,445 INFO Thread-12 :30410 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_190333-82mnef5m/files/output.log +2024-06-08 19:45:33,849 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 19:45:34,269 DEBUG SenderThread:30410 [sender.py:send():379] send: stats +2024-06-08 19:45:34,383 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: stop_status +2024-06-08 19:45:34,383 DEBUG SenderThread:30410 [sender.py:send_request():406] send_request: stop_status +2024-06-08 19:45:35,483 INFO Thread-12 :30410 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_190333-82mnef5m/files/output.log +2024-06-08 19:45:38,215 INFO Thread-12 :30410 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_190333-82mnef5m/files/output.log +2024-06-08 19:45:38,850 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 19:45:43,857 INFO Thread-12 :30410 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_190333-82mnef5m/files/output.log +2024-06-08 19:45:44,585 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 19:45:49,383 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: stop_status +2024-06-08 19:45:49,383 DEBUG SenderThread:30410 [sender.py:send_request():406] send_request: stop_status +2024-06-08 19:45:50,524 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 19:45:52,127 INFO Thread-12 :30410 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_190333-82mnef5m/files/output.log +2024-06-08 19:45:55,850 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 19:45:58,011 INFO Thread-12 :30410 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_190333-82mnef5m/files/output.log +2024-06-08 19:46:00,850 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 19:46:04,271 DEBUG SenderThread:30410 [sender.py:send():379] send: stats +2024-06-08 19:46:04,383 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: stop_status +2024-06-08 19:46:04,383 DEBUG SenderThread:30410 [sender.py:send_request():406] send_request: stop_status +2024-06-08 19:46:05,850 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 19:46:06,331 INFO Thread-12 :30410 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_190333-82mnef5m/files/output.log +2024-06-08 19:46:11,850 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 19:46:11,987 INFO Thread-12 :30410 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_190333-82mnef5m/files/output.log +2024-06-08 19:46:16,850 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 19:46:19,383 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: stop_status +2024-06-08 19:46:19,383 DEBUG SenderThread:30410 [sender.py:send_request():406] send_request: stop_status +2024-06-08 19:46:20,307 INFO Thread-12 :30410 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_190333-82mnef5m/files/output.log +2024-06-08 19:46:22,850 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 19:46:26,058 INFO Thread-12 :30410 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_190333-82mnef5m/files/output.log +2024-06-08 19:46:27,850 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 19:46:33,738 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 19:46:34,273 DEBUG SenderThread:30410 [sender.py:send():379] send: stats +2024-06-08 19:46:34,290 INFO Thread-12 :30410 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_190333-82mnef5m/files/output.log +2024-06-08 19:46:34,383 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: stop_status +2024-06-08 19:46:34,383 DEBUG SenderThread:30410 [sender.py:send_request():406] send_request: stop_status +2024-06-08 19:46:39,700 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 19:46:40,128 INFO Thread-12 :30410 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_190333-82mnef5m/files/output.log +2024-06-08 19:46:44,851 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 19:46:48,211 INFO Thread-12 :30410 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_190333-82mnef5m/files/output.log +2024-06-08 19:46:49,383 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: stop_status +2024-06-08 19:46:49,384 DEBUG SenderThread:30410 [sender.py:send_request():406] send_request: stop_status +2024-06-08 19:46:50,469 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 19:46:53,884 INFO Thread-12 :30410 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_190333-82mnef5m/files/output.log +2024-06-08 19:46:55,851 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 19:46:59,608 INFO Thread-12 :30410 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_190333-82mnef5m/files/output.log +2024-06-08 19:47:00,851 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 19:47:01,851 INFO Thread-12 :30410 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_190333-82mnef5m/files/output.log +2024-06-08 19:47:04,276 DEBUG SenderThread:30410 [sender.py:send():379] send: stats +2024-06-08 19:47:04,383 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: stop_status +2024-06-08 19:47:04,384 DEBUG SenderThread:30410 [sender.py:send_request():406] send_request: stop_status +2024-06-08 19:47:05,851 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 19:47:07,878 INFO Thread-12 :30410 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_190333-82mnef5m/files/output.log +2024-06-08 19:47:11,851 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 19:47:16,123 INFO Thread-12 :30410 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_190333-82mnef5m/files/output.log +2024-06-08 19:47:17,851 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 19:47:19,384 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: stop_status +2024-06-08 19:47:19,384 DEBUG SenderThread:30410 [sender.py:send_request():406] send_request: stop_status +2024-06-08 19:47:21,676 INFO Thread-12 :30410 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_190333-82mnef5m/files/output.log +2024-06-08 19:47:23,851 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 19:47:29,349 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 19:47:29,771 INFO Thread-12 :30410 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_190333-82mnef5m/files/output.log +2024-06-08 19:47:34,278 DEBUG SenderThread:30410 [sender.py:send():379] send: stats +2024-06-08 19:47:34,384 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: stop_status +2024-06-08 19:47:34,384 DEBUG SenderThread:30410 [sender.py:send_request():406] send_request: stop_status +2024-06-08 19:47:34,473 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 19:47:35,618 INFO Thread-12 :30410 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_190333-82mnef5m/files/output.log +2024-06-08 19:47:39,852 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 19:47:41,459 INFO Thread-12 :30410 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_190333-82mnef5m/files/output.log +2024-06-08 19:47:43,843 INFO Thread-12 :30410 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_190333-82mnef5m/files/output.log +2024-06-08 19:47:45,208 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 19:47:49,384 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: stop_status +2024-06-08 19:47:49,384 DEBUG SenderThread:30410 [sender.py:send_request():406] send_request: stop_status +2024-06-08 19:47:49,524 INFO Thread-12 :30410 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_190333-82mnef5m/files/output.log +2024-06-08 19:47:50,524 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 19:47:55,423 INFO Thread-12 :30410 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_190333-82mnef5m/files/output.log +2024-06-08 19:47:55,852 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 19:47:57,799 INFO Thread-12 :30410 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_190333-82mnef5m/files/output.log +2024-06-08 19:48:00,852 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 19:48:03,795 INFO Thread-12 :30410 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_190333-82mnef5m/files/output.log +2024-06-08 19:48:04,280 DEBUG SenderThread:30410 [sender.py:send():379] send: stats +2024-06-08 19:48:04,384 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: stop_status +2024-06-08 19:48:04,384 DEBUG SenderThread:30410 [sender.py:send_request():406] send_request: stop_status +2024-06-08 19:48:05,852 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 19:48:09,631 INFO Thread-12 :30410 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_190333-82mnef5m/files/output.log +2024-06-08 19:48:10,852 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 19:48:15,852 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 19:48:17,881 INFO Thread-12 :30410 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_190333-82mnef5m/files/output.log +2024-06-08 19:48:19,384 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: stop_status +2024-06-08 19:48:19,384 DEBUG SenderThread:30410 [sender.py:send_request():406] send_request: stop_status +2024-06-08 19:48:20,852 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 19:48:23,483 INFO Thread-12 :30410 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_190333-82mnef5m/files/output.log +2024-06-08 19:48:25,853 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 19:48:30,322 INFO Thread-12 :30410 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_190333-82mnef5m/files/output.log +2024-06-08 19:48:30,853 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 19:48:31,568 INFO Thread-12 :30410 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_190333-82mnef5m/files/output.log +2024-06-08 19:48:34,282 DEBUG SenderThread:30410 [sender.py:send():379] send: stats +2024-06-08 19:48:34,384 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: stop_status +2024-06-08 19:48:34,384 DEBUG SenderThread:30410 [sender.py:send_request():406] send_request: stop_status +2024-06-08 19:48:35,915 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 19:48:38,579 INFO Thread-12 :30410 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_190333-82mnef5m/files/output.log +2024-06-08 19:48:41,853 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 19:48:44,186 INFO Thread-12 :30410 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_190333-82mnef5m/files/output.log +2024-06-08 19:48:47,853 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 19:48:49,384 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: stop_status +2024-06-08 19:48:49,384 DEBUG SenderThread:30410 [sender.py:send_request():406] send_request: stop_status +2024-06-08 19:48:51,900 INFO Thread-12 :30410 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_190333-82mnef5m/files/output.log +2024-06-08 19:48:52,853 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 19:48:57,864 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 19:48:57,883 INFO Thread-12 :30410 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_190333-82mnef5m/files/output.log +2024-06-08 19:49:00,200 INFO Thread-12 :30410 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_190333-82mnef5m/files/output.log +2024-06-08 19:49:03,320 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 19:49:04,284 DEBUG SenderThread:30410 [sender.py:send():379] send: stats +2024-06-08 19:49:04,384 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: stop_status +2024-06-08 19:49:04,384 DEBUG SenderThread:30410 [sender.py:send_request():406] send_request: stop_status +2024-06-08 19:49:06,107 INFO Thread-12 :30410 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_190333-82mnef5m/files/output.log +2024-06-08 19:49:08,853 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 19:49:11,787 INFO Thread-12 :30410 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_190333-82mnef5m/files/output.log +2024-06-08 19:49:14,853 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 19:49:19,384 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: stop_status +2024-06-08 19:49:19,385 DEBUG SenderThread:30410 [sender.py:send_request():406] send_request: stop_status +2024-06-08 19:49:19,987 INFO Thread-12 :30410 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_190333-82mnef5m/files/output.log +2024-06-08 19:49:20,542 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 19:49:25,673 INFO Thread-12 :30410 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_190333-82mnef5m/files/output.log +2024-06-08 19:49:25,854 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 19:49:30,874 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 19:49:31,411 INFO Thread-12 :30410 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_190333-82mnef5m/files/output.log +2024-06-08 19:49:34,286 DEBUG SenderThread:30410 [sender.py:send():379] send: stats +2024-06-08 19:49:34,384 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: stop_status +2024-06-08 19:49:34,385 DEBUG SenderThread:30410 [sender.py:send_request():406] send_request: stop_status +2024-06-08 19:49:36,854 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 19:49:39,451 INFO Thread-12 :30410 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_190333-82mnef5m/files/output.log +2024-06-08 19:49:41,854 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 19:49:46,491 INFO Thread-12 :30410 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_190333-82mnef5m/files/output.log +2024-06-08 19:49:47,787 INFO Thread-12 :30410 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_190333-82mnef5m/files/output.log +2024-06-08 19:49:47,854 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 19:49:49,384 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: stop_status +2024-06-08 19:49:49,385 DEBUG SenderThread:30410 [sender.py:send_request():406] send_request: stop_status +2024-06-08 19:49:53,231 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 19:49:53,535 INFO Thread-12 :30410 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_190333-82mnef5m/files/output.log +2024-06-08 19:49:58,771 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 19:50:00,236 INFO Thread-12 :30410 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_190333-82mnef5m/files/output.log +2024-06-08 19:50:01,555 INFO Thread-12 :30410 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_190333-82mnef5m/files/output.log +2024-06-08 19:50:03,854 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 19:50:04,288 DEBUG SenderThread:30410 [sender.py:send():379] send: stats +2024-06-08 19:50:04,384 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: stop_status +2024-06-08 19:50:04,385 DEBUG SenderThread:30410 [sender.py:send_request():406] send_request: stop_status +2024-06-08 19:50:08,371 INFO Thread-12 :30410 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_190333-82mnef5m/files/output.log +2024-06-08 19:50:08,884 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 19:50:14,300 INFO Thread-12 :30410 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_190333-82mnef5m/files/output.log +2024-06-08 19:50:14,837 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 19:50:16,679 INFO Thread-12 :30410 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_190333-82mnef5m/files/output.log +2024-06-08 19:50:19,385 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: stop_status +2024-06-08 19:50:19,385 DEBUG SenderThread:30410 [sender.py:send_request():406] send_request: stop_status +2024-06-08 19:50:20,453 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 19:50:22,335 INFO Thread-12 :30410 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_190333-82mnef5m/files/output.log +2024-06-08 19:50:24,532 INFO Thread-12 :30410 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_190333-82mnef5m/files/output.log +2024-06-08 19:50:25,855 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 19:50:30,472 INFO Thread-12 :30410 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_190333-82mnef5m/files/output.log +2024-06-08 19:50:30,855 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 19:50:34,291 DEBUG SenderThread:30410 [sender.py:send():379] send: stats +2024-06-08 19:50:34,385 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: stop_status +2024-06-08 19:50:34,385 DEBUG SenderThread:30410 [sender.py:send_request():406] send_request: stop_status +2024-06-08 19:50:36,214 INFO Thread-12 :30410 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_190333-82mnef5m/files/output.log +2024-06-08 19:50:36,855 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 19:50:41,878 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 19:50:44,579 INFO Thread-12 :30410 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_190333-82mnef5m/files/output.log +2024-06-08 19:50:47,855 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 19:50:49,385 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: stop_status +2024-06-08 19:50:49,385 DEBUG SenderThread:30410 [sender.py:send_request():406] send_request: stop_status +2024-06-08 19:50:50,174 INFO Thread-12 :30410 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_190333-82mnef5m/files/output.log +2024-06-08 19:50:53,855 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 19:50:57,907 INFO Thread-12 :30410 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_190333-82mnef5m/files/output.log +2024-06-08 19:50:58,855 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 19:51:03,455 INFO Thread-12 :30410 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_190333-82mnef5m/files/output.log +2024-06-08 19:51:03,893 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 19:51:04,294 DEBUG SenderThread:30410 [sender.py:send():379] send: stats +2024-06-08 19:51:04,385 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: stop_status +2024-06-08 19:51:04,385 DEBUG SenderThread:30410 [sender.py:send_request():406] send_request: stop_status +2024-06-08 19:51:09,869 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 19:51:10,262 INFO Thread-12 :30410 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_190333-82mnef5m/files/output.log +2024-06-08 19:51:15,182 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 19:51:18,083 INFO Thread-12 :30410 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_190333-82mnef5m/files/output.log +2024-06-08 19:51:19,385 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: stop_status +2024-06-08 19:51:19,385 DEBUG SenderThread:30410 [sender.py:send_request():406] send_request: stop_status +2024-06-08 19:51:20,551 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 19:51:23,776 INFO Thread-12 :30410 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_190333-82mnef5m/files/output.log +2024-06-08 19:51:25,856 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 19:51:30,081 INFO Thread-12 :30410 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_190333-82mnef5m/files/output.log +2024-06-08 19:51:30,856 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 19:51:32,448 INFO Thread-12 :30410 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_190333-82mnef5m/files/output.log +2024-06-08 19:51:34,295 DEBUG SenderThread:30410 [sender.py:send():379] send: stats +2024-06-08 19:51:34,385 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: stop_status +2024-06-08 19:51:34,385 DEBUG SenderThread:30410 [sender.py:send_request():406] send_request: stop_status +2024-06-08 19:51:35,856 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 19:51:38,468 INFO Thread-12 :30410 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_190333-82mnef5m/files/output.log +2024-06-08 19:51:40,856 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 19:51:44,275 INFO Thread-12 :30410 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_190333-82mnef5m/files/output.log +2024-06-08 19:51:45,856 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 19:51:49,414 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: stop_status +2024-06-08 19:51:49,414 DEBUG SenderThread:30410 [sender.py:send_request():406] send_request: stop_status +2024-06-08 19:51:51,856 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 19:51:51,891 INFO Thread-12 :30410 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_190333-82mnef5m/files/output.log +2024-06-08 19:51:57,540 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 19:51:57,866 INFO Thread-12 :30410 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_190333-82mnef5m/files/output.log +2024-06-08 19:52:02,857 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 19:52:04,297 DEBUG SenderThread:30410 [sender.py:send():379] send: stats +2024-06-08 19:52:04,385 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: stop_status +2024-06-08 19:52:04,385 DEBUG SenderThread:30410 [sender.py:send_request():406] send_request: stop_status +2024-06-08 19:52:04,583 INFO Thread-12 :30410 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_190333-82mnef5m/files/output.log +2024-06-08 19:52:05,907 INFO Thread-12 :30410 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_190333-82mnef5m/files/output.log +2024-06-08 19:52:08,857 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 19:52:11,658 INFO Thread-12 :30410 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_190333-82mnef5m/files/output.log +2024-06-08 19:52:14,857 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 19:52:17,811 INFO Thread-12 :30410 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_190333-82mnef5m/files/output.log +2024-06-08 19:52:19,385 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: stop_status +2024-06-08 19:52:19,386 DEBUG SenderThread:30410 [sender.py:send_request():406] send_request: stop_status +2024-06-08 19:52:20,175 INFO Thread-12 :30410 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_190333-82mnef5m/files/output.log +2024-06-08 19:52:20,488 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 19:52:25,857 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 19:52:26,003 INFO Thread-12 :30410 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_190333-82mnef5m/files/output.log +2024-06-08 19:52:30,857 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 19:52:31,741 INFO Thread-12 :30410 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_190333-82mnef5m/files/output.log +2024-06-08 19:52:34,299 DEBUG SenderThread:30410 [sender.py:send():379] send: stats +2024-06-08 19:52:34,385 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: stop_status +2024-06-08 19:52:34,386 DEBUG SenderThread:30410 [sender.py:send_request():406] send_request: stop_status +2024-06-08 19:52:35,857 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 19:52:37,747 INFO Thread-12 :30410 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_190333-82mnef5m/files/output.log +2024-06-08 19:52:40,857 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 19:52:46,251 INFO Thread-12 :30410 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_190333-82mnef5m/files/output.log +2024-06-08 19:52:46,857 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 19:52:49,385 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: stop_status +2024-06-08 19:52:49,386 DEBUG SenderThread:30410 [sender.py:send_request():406] send_request: stop_status +2024-06-08 19:52:51,923 INFO Thread-12 :30410 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_190333-82mnef5m/files/output.log +2024-06-08 19:52:52,432 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 19:52:53,987 INFO Thread-12 :30410 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_190333-82mnef5m/files/output.log +2024-06-08 19:52:57,803 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 19:52:59,864 INFO Thread-12 :30410 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_190333-82mnef5m/files/output.log +2024-06-08 19:53:02,858 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 19:53:04,302 DEBUG SenderThread:30410 [sender.py:send():379] send: stats +2024-06-08 19:53:04,385 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: stop_status +2024-06-08 19:53:04,386 DEBUG SenderThread:30410 [sender.py:send_request():406] send_request: stop_status +2024-06-08 19:53:05,716 INFO Thread-12 :30410 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_190333-82mnef5m/files/output.log +2024-06-08 19:53:08,858 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 19:53:13,595 INFO Thread-12 :30410 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_190333-82mnef5m/files/output.log +2024-06-08 19:53:14,483 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 19:53:19,386 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: stop_status +2024-06-08 19:53:19,386 DEBUG SenderThread:30410 [sender.py:send_request():406] send_request: stop_status +2024-06-08 19:53:20,299 INFO Thread-12 :30410 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_190333-82mnef5m/files/output.log +2024-06-08 19:53:20,464 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 19:53:25,858 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 19:53:26,015 INFO Thread-12 :30410 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_190333-82mnef5m/files/output.log +2024-06-08 19:53:30,858 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 19:53:33,943 INFO Thread-12 :30410 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_190333-82mnef5m/files/output.log +2024-06-08 19:53:34,303 DEBUG SenderThread:30410 [sender.py:send():379] send: stats +2024-06-08 19:53:34,386 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: stop_status +2024-06-08 19:53:34,386 DEBUG SenderThread:30410 [sender.py:send_request():406] send_request: stop_status +2024-06-08 19:53:35,858 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 19:53:39,803 INFO Thread-12 :30410 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_190333-82mnef5m/files/output.log +2024-06-08 19:53:40,859 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 19:53:46,593 INFO Thread-12 :30410 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_190333-82mnef5m/files/output.log +2024-06-08 19:53:46,860 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 19:53:49,386 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: stop_status +2024-06-08 19:53:49,386 DEBUG SenderThread:30410 [sender.py:send_request():406] send_request: stop_status +2024-06-08 19:53:52,666 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 19:53:54,231 INFO Thread-12 :30410 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_190333-82mnef5m/files/output.log +2024-06-08 19:53:57,859 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 19:54:00,212 INFO Thread-12 :30410 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_190333-82mnef5m/files/output.log +2024-06-08 19:54:03,859 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 19:54:04,306 DEBUG SenderThread:30410 [sender.py:send():379] send: stats +2024-06-08 19:54:04,386 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: stop_status +2024-06-08 19:54:04,386 DEBUG SenderThread:30410 [sender.py:send_request():406] send_request: stop_status +2024-06-08 19:54:08,611 INFO Thread-12 :30410 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_190333-82mnef5m/files/output.log +2024-06-08 19:54:09,461 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 19:54:14,324 INFO Thread-12 :30410 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_190333-82mnef5m/files/output.log +2024-06-08 19:54:15,409 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 19:54:16,803 INFO Thread-12 :30410 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_190333-82mnef5m/files/output.log +2024-06-08 19:54:19,386 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: stop_status +2024-06-08 19:54:19,387 DEBUG SenderThread:30410 [sender.py:send_request():406] send_request: stop_status +2024-06-08 19:54:20,859 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 19:54:22,748 INFO Thread-12 :30410 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_190333-82mnef5m/files/output.log +2024-06-08 19:54:25,859 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 19:54:28,548 INFO Thread-12 :30410 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_190333-82mnef5m/files/output.log +2024-06-08 19:54:30,859 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 19:54:34,308 DEBUG SenderThread:30410 [sender.py:send():379] send: stats +2024-06-08 19:54:34,386 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: stop_status +2024-06-08 19:54:34,387 DEBUG SenderThread:30410 [sender.py:send_request():406] send_request: stop_status +2024-06-08 19:54:34,555 INFO Thread-12 :30410 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_190333-82mnef5m/files/output.log +2024-06-08 19:54:36,839 INFO Thread-12 :30410 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_190333-82mnef5m/files/output.log +2024-06-08 19:54:36,859 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 19:54:42,237 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 19:54:42,588 INFO Thread-12 :30410 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_190333-82mnef5m/files/output.log +2024-06-08 19:54:47,627 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 19:54:49,387 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: stop_status +2024-06-08 19:54:49,387 DEBUG SenderThread:30410 [sender.py:send_request():406] send_request: stop_status +2024-06-08 19:54:50,587 INFO Thread-12 :30410 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_190333-82mnef5m/files/output.log +2024-06-08 19:54:52,860 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 19:54:56,464 INFO Thread-12 :30410 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_190333-82mnef5m/files/output.log +2024-06-08 19:54:58,860 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 19:55:04,132 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 19:55:04,310 DEBUG SenderThread:30410 [sender.py:send():379] send: stats +2024-06-08 19:55:04,387 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: stop_status +2024-06-08 19:55:04,387 DEBUG SenderThread:30410 [sender.py:send_request():406] send_request: stop_status +2024-06-08 19:55:04,627 INFO Thread-12 :30410 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_190333-82mnef5m/files/output.log +2024-06-08 19:55:09,422 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 19:55:10,164 INFO Thread-12 :30410 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_190333-82mnef5m/files/output.log +2024-06-08 19:55:14,618 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 19:55:15,859 INFO Thread-12 :30410 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_190333-82mnef5m/files/output.log +2024-06-08 19:55:19,387 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: stop_status +2024-06-08 19:55:19,387 DEBUG SenderThread:30410 [sender.py:send_request():406] send_request: stop_status +2024-06-08 19:55:20,451 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 19:55:21,767 INFO Thread-12 :30410 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_190333-82mnef5m/files/output.log +2024-06-08 19:55:23,967 INFO Thread-12 :30410 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_190333-82mnef5m/files/output.log +2024-06-08 19:55:25,853 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 19:55:30,566 INFO Thread-12 :30410 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_190333-82mnef5m/files/output.log +2024-06-08 19:55:30,860 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 19:55:34,312 DEBUG SenderThread:30410 [sender.py:send():379] send: stats +2024-06-08 19:55:34,387 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: stop_status +2024-06-08 19:55:34,387 DEBUG SenderThread:30410 [sender.py:send_request():406] send_request: stop_status +2024-06-08 19:55:35,860 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 19:55:36,522 INFO Thread-12 :30410 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_190333-82mnef5m/files/output.log +2024-06-08 19:55:40,860 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 19:55:42,086 INFO Thread-12 :30410 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_190333-82mnef5m/files/output.log +2024-06-08 19:55:44,364 INFO Thread-12 :30410 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_190333-82mnef5m/files/output.log +2024-06-08 19:55:45,861 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 19:55:49,387 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: stop_status +2024-06-08 19:55:49,387 DEBUG SenderThread:30410 [sender.py:send_request():406] send_request: stop_status +2024-06-08 19:55:50,263 INFO Thread-12 :30410 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_190333-82mnef5m/files/output.log +2024-06-08 19:55:50,861 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 19:55:55,861 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 19:55:56,443 INFO Thread-12 :30410 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_190333-82mnef5m/files/output.log +2024-06-08 19:56:00,861 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 19:56:02,135 INFO Thread-12 :30410 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_190333-82mnef5m/files/output.log +2024-06-08 19:56:04,314 DEBUG SenderThread:30410 [sender.py:send():379] send: stats +2024-06-08 19:56:04,387 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: stop_status +2024-06-08 19:56:04,387 DEBUG SenderThread:30410 [sender.py:send_request():406] send_request: stop_status +2024-06-08 19:56:04,531 INFO Thread-12 :30410 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_190333-82mnef5m/files/output.log +2024-06-08 19:56:06,861 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 19:56:10,547 INFO Thread-12 :30410 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_190333-82mnef5m/files/output.log +2024-06-08 19:56:12,861 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 19:56:16,399 INFO Thread-12 :30410 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_190333-82mnef5m/files/output.log +2024-06-08 19:56:18,861 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 19:56:19,387 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: stop_status +2024-06-08 19:56:19,387 DEBUG SenderThread:30410 [sender.py:send_request():406] send_request: stop_status +2024-06-08 19:56:22,267 INFO Thread-12 :30410 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_190333-82mnef5m/files/output.log +2024-06-08 19:56:24,563 INFO Thread-12 :30410 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_190333-82mnef5m/files/output.log +2024-06-08 19:56:24,676 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 19:56:29,929 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 19:56:30,138 INFO Thread-12 :30410 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_190333-82mnef5m/files/output.log +2024-06-08 19:56:34,316 DEBUG SenderThread:30410 [sender.py:send():379] send: stats +2024-06-08 19:56:34,387 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: stop_status +2024-06-08 19:56:34,387 DEBUG SenderThread:30410 [sender.py:send_request():406] send_request: stop_status +2024-06-08 19:56:35,543 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 19:56:36,952 INFO Thread-12 :30410 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_190333-82mnef5m/files/output.log +2024-06-08 19:56:40,552 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 19:56:42,594 INFO Thread-12 :30410 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_190333-82mnef5m/files/output.log +2024-06-08 19:56:45,864 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 19:56:48,339 INFO Thread-12 :30410 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_190333-82mnef5m/files/output.log +2024-06-08 19:56:49,387 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: stop_status +2024-06-08 19:56:49,388 DEBUG SenderThread:30410 [sender.py:send_request():406] send_request: stop_status +2024-06-08 19:56:50,495 INFO Thread-12 :30410 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_190333-82mnef5m/files/output.log +2024-06-08 19:56:51,864 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 19:56:56,564 INFO Thread-12 :30410 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_190333-82mnef5m/files/output.log +2024-06-08 19:56:56,864 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 19:57:01,907 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 19:57:02,346 INFO Thread-12 :30410 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_190333-82mnef5m/files/output.log +2024-06-08 19:57:04,318 DEBUG SenderThread:30410 [sender.py:send():379] send: stats +2024-06-08 19:57:04,387 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: stop_status +2024-06-08 19:57:04,388 DEBUG SenderThread:30410 [sender.py:send_request():406] send_request: stop_status +2024-06-08 19:57:07,190 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 19:57:08,175 INFO Thread-12 :30410 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_190333-82mnef5m/files/output.log +2024-06-08 19:57:12,864 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 19:57:16,572 INFO Thread-12 :30410 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_190333-82mnef5m/files/output.log +2024-06-08 19:57:17,864 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 19:57:19,387 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: stop_status +2024-06-08 19:57:19,388 DEBUG SenderThread:30410 [sender.py:send_request():406] send_request: stop_status +2024-06-08 19:57:22,455 INFO Thread-12 :30410 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_190333-82mnef5m/files/output.log +2024-06-08 19:57:23,149 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 19:57:28,587 INFO Thread-12 :30410 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_190333-82mnef5m/files/output.log +2024-06-08 19:57:28,643 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 19:57:29,959 INFO Thread-12 :30410 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_190333-82mnef5m/files/output.log +2024-06-08 19:57:33,865 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 19:57:34,320 DEBUG SenderThread:30410 [sender.py:send():379] send: stats +2024-06-08 19:57:34,387 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: stop_status +2024-06-08 19:57:34,388 DEBUG SenderThread:30410 [sender.py:send_request():406] send_request: stop_status +2024-06-08 19:57:37,144 INFO Thread-12 :30410 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_190333-82mnef5m/files/output.log +2024-06-08 19:57:38,874 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 19:57:42,023 INFO Thread-12 :30410 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_190333-82mnef5m/files/output.log +2024-06-08 19:57:44,134 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 19:57:49,084 INFO Thread-12 :30410 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_190333-82mnef5m/files/output.log +2024-06-08 19:57:49,388 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: stop_status +2024-06-08 19:57:49,388 DEBUG SenderThread:30410 [sender.py:send_request():406] send_request: stop_status +2024-06-08 19:57:49,556 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 19:57:50,331 INFO Thread-12 :30410 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_190333-82mnef5m/files/output.log +2024-06-08 19:57:55,386 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 19:57:56,069 INFO Thread-12 :30410 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_190333-82mnef5m/files/output.log +2024-06-08 19:58:00,579 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 19:58:03,012 INFO Thread-12 :30410 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_190333-82mnef5m/files/output.log +2024-06-08 19:58:04,231 INFO Thread-12 :30410 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_190333-82mnef5m/files/output.log +2024-06-08 19:58:04,322 DEBUG SenderThread:30410 [sender.py:send():379] send: stats +2024-06-08 19:58:04,388 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: stop_status +2024-06-08 19:58:04,388 DEBUG SenderThread:30410 [sender.py:send_request():406] send_request: stop_status +2024-06-08 19:58:05,865 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 19:58:10,865 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 19:58:11,108 INFO Thread-12 :30410 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_190333-82mnef5m/files/output.log +2024-06-08 19:58:15,865 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 19:58:17,160 INFO Thread-12 :30410 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_190333-82mnef5m/files/output.log +2024-06-08 19:58:19,388 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: stop_status +2024-06-08 19:58:19,388 DEBUG SenderThread:30410 [sender.py:send_request():406] send_request: stop_status +2024-06-08 19:58:20,866 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 19:58:22,851 INFO Thread-12 :30410 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_190333-82mnef5m/files/output.log +2024-06-08 19:58:24,022 INFO Thread-12 :30410 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_190333-82mnef5m/files/output.log +2024-06-08 19:58:25,866 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 19:58:30,866 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 19:58:30,931 INFO Thread-12 :30410 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_190333-82mnef5m/files/output.log +2024-06-08 19:58:34,324 DEBUG SenderThread:30410 [sender.py:send():379] send: stats +2024-06-08 19:58:34,388 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: stop_status +2024-06-08 19:58:34,388 DEBUG SenderThread:30410 [sender.py:send_request():406] send_request: stop_status +2024-06-08 19:58:35,866 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 19:58:37,172 INFO Thread-12 :30410 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_190333-82mnef5m/files/output.log +2024-06-08 19:58:41,866 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 19:58:43,112 INFO Thread-12 :30410 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_190333-82mnef5m/files/output.log +2024-06-08 19:58:47,585 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 19:58:49,116 INFO Thread-12 :30410 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_190333-82mnef5m/files/output.log +2024-06-08 19:58:49,388 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: stop_status +2024-06-08 19:58:49,388 DEBUG SenderThread:30410 [sender.py:send_request():406] send_request: stop_status +2024-06-08 19:58:53,051 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 19:58:54,643 INFO Thread-12 :30410 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_190333-82mnef5m/files/output.log +2024-06-08 19:58:58,011 INFO Thread-12 :30410 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_190333-82mnef5m/files/output.log +2024-06-08 19:58:58,536 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 19:59:04,075 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 19:59:04,326 DEBUG SenderThread:30410 [sender.py:send():379] send: stats +2024-06-08 19:59:04,388 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: stop_status +2024-06-08 19:59:04,388 DEBUG SenderThread:30410 [sender.py:send_request():406] send_request: stop_status +2024-06-08 19:59:05,320 INFO Thread-12 :30410 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_190333-82mnef5m/files/output.log +2024-06-08 19:59:10,069 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 19:59:11,200 INFO Thread-12 :30410 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_190333-82mnef5m/files/output.log +2024-06-08 19:59:15,867 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 19:59:17,148 INFO Thread-12 :30410 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_190333-82mnef5m/files/output.log +2024-06-08 19:59:19,388 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: stop_status +2024-06-08 19:59:19,388 DEBUG SenderThread:30410 [sender.py:send_request():406] send_request: stop_status +2024-06-08 19:59:21,867 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 19:59:23,076 INFO Thread-12 :30410 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_190333-82mnef5m/files/output.log +2024-06-08 19:59:27,259 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 19:59:29,084 INFO Thread-12 :30410 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_190333-82mnef5m/files/output.log +2024-06-08 19:59:32,668 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 19:59:32,768 INFO Thread-12 :30410 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_190333-82mnef5m/files/output.log +2024-06-08 19:59:34,328 DEBUG SenderThread:30410 [sender.py:send():379] send: stats +2024-06-08 19:59:34,388 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: stop_status +2024-06-08 19:59:34,388 DEBUG SenderThread:30410 [sender.py:send_request():406] send_request: stop_status +2024-06-08 19:59:37,728 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 19:59:38,347 INFO Thread-12 :30410 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_190333-82mnef5m/files/output.log +2024-06-08 19:59:40,648 INFO Thread-12 :30410 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_190333-82mnef5m/files/output.log +2024-06-08 19:59:42,800 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 19:59:45,128 INFO Thread-12 :30410 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_190333-82mnef5m/files/output.log +2024-06-08 19:59:47,867 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 19:59:49,388 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: stop_status +2024-06-08 19:59:49,389 DEBUG SenderThread:30410 [sender.py:send_request():406] send_request: stop_status +2024-06-08 19:59:50,855 INFO Thread-12 :30410 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_190333-82mnef5m/files/output.log +2024-06-08 19:59:53,372 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 19:59:54,643 INFO Thread-12 :30410 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_190333-82mnef5m/files/output.log +2024-06-08 19:59:57,048 INFO Thread-12 :30410 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_190333-82mnef5m/files/output.log +2024-06-08 19:59:58,823 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 20:00:00,324 INFO Thread-12 :30410 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_190333-82mnef5m/files/output.log +2024-06-08 20:00:03,867 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 20:00:04,331 DEBUG SenderThread:30410 [sender.py:send():379] send: stats +2024-06-08 20:00:04,388 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: stop_status +2024-06-08 20:00:04,389 DEBUG SenderThread:30410 [sender.py:send_request():406] send_request: stop_status +2024-06-08 20:00:07,004 INFO Thread-12 :30410 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_190333-82mnef5m/files/output.log +2024-06-08 20:00:08,868 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 20:00:12,716 INFO Thread-12 :30410 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_190333-82mnef5m/files/output.log +2024-06-08 20:00:14,868 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 20:00:16,223 INFO Thread-12 :30410 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_190333-82mnef5m/files/output.log +2024-06-08 20:00:18,512 INFO Thread-12 :30410 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_190333-82mnef5m/files/output.log +2024-06-08 20:00:19,389 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: stop_status +2024-06-08 20:00:19,389 DEBUG SenderThread:30410 [sender.py:send_request():406] send_request: stop_status +2024-06-08 20:00:20,544 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 20:00:23,208 INFO Thread-12 :30410 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_190333-82mnef5m/files/output.log +2024-06-08 20:00:24,307 INFO Thread-12 :30410 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_190333-82mnef5m/files/output.log +2024-06-08 20:00:25,868 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 20:00:29,027 INFO Thread-12 :30410 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_190333-82mnef5m/files/output.log +2024-06-08 20:00:30,868 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 20:00:34,333 DEBUG SenderThread:30410 [sender.py:send():379] send: stats +2024-06-08 20:00:34,389 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: stop_status +2024-06-08 20:00:34,389 DEBUG SenderThread:30410 [sender.py:send_request():406] send_request: stop_status +2024-06-08 20:00:34,400 INFO Thread-12 :30410 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_190333-82mnef5m/files/output.log +2024-06-08 20:00:36,868 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 20:00:41,383 INFO Thread-12 :30410 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_190333-82mnef5m/files/output.log +2024-06-08 20:00:41,868 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 20:00:44,883 INFO Thread-12 :30410 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_190333-82mnef5m/files/output.log +2024-06-08 20:00:47,188 INFO Thread-12 :30410 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_190333-82mnef5m/files/output.log +2024-06-08 20:00:47,868 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 20:00:49,389 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: stop_status +2024-06-08 20:00:49,389 DEBUG SenderThread:30410 [sender.py:send_request():406] send_request: stop_status +2024-06-08 20:00:50,523 INFO Thread-12 :30410 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_190333-82mnef5m/files/output.log +2024-06-08 20:00:53,868 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 20:00:57,285 INFO Thread-12 :30410 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_190333-82mnef5m/files/output.log +2024-06-08 20:00:58,869 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 20:01:00,567 INFO Thread-12 :30410 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_190333-82mnef5m/files/output.log +2024-06-08 20:01:03,076 INFO Thread-12 :30410 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_190333-82mnef5m/files/output.log +2024-06-08 20:01:03,923 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 20:01:04,335 DEBUG SenderThread:30410 [sender.py:send():379] send: stats +2024-06-08 20:01:04,389 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: stop_status +2024-06-08 20:01:04,389 DEBUG SenderThread:30410 [sender.py:send_request():406] send_request: stop_status +2024-06-08 20:01:08,692 INFO Thread-12 :30410 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_190333-82mnef5m/files/output.log +2024-06-08 20:01:09,071 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 20:01:12,456 INFO Thread-12 :30410 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_190333-82mnef5m/files/output.log +2024-06-08 20:01:14,505 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 20:01:18,379 INFO Thread-12 :30410 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_190333-82mnef5m/files/output.log +2024-06-08 20:01:19,389 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: stop_status +2024-06-08 20:01:19,389 DEBUG SenderThread:30410 [sender.py:send_request():406] send_request: stop_status +2024-06-08 20:01:19,518 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 20:01:24,451 INFO Thread-12 :30410 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_190333-82mnef5m/files/output.log +2024-06-08 20:01:25,264 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 20:01:29,025 INFO Thread-12 :30410 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_190333-82mnef5m/files/output.log +2024-06-08 20:01:30,652 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 20:01:34,346 DEBUG SenderThread:30410 [sender.py:send():379] send: stats +2024-06-08 20:01:34,389 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: stop_status +2024-06-08 20:01:34,389 DEBUG SenderThread:30410 [sender.py:send_request():406] send_request: stop_status +2024-06-08 20:01:34,867 INFO Thread-12 :30410 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_190333-82mnef5m/files/output.log +2024-06-08 20:01:35,869 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 20:01:38,311 INFO Thread-12 :30410 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_190333-82mnef5m/files/output.log +2024-06-08 20:01:40,408 INFO Thread-12 :30410 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_190333-82mnef5m/files/output.log +2024-06-08 20:01:40,869 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 20:01:45,452 INFO Thread-12 :30410 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_190333-82mnef5m/files/output.log +2024-06-08 20:01:45,869 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 20:01:49,389 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: stop_status +2024-06-08 20:01:49,389 DEBUG SenderThread:30410 [sender.py:send_request():406] send_request: stop_status +2024-06-08 20:01:50,870 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 20:01:51,156 INFO Thread-12 :30410 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_190333-82mnef5m/files/output.log +2024-06-08 20:01:54,703 INFO Thread-12 :30410 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_190333-82mnef5m/files/output.log +2024-06-08 20:01:56,870 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 20:02:00,747 INFO Thread-12 :30410 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_190333-82mnef5m/files/output.log +2024-06-08 20:02:01,870 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 20:02:02,835 INFO Thread-12 :30410 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_190333-82mnef5m/files/output.log +2024-06-08 20:02:04,339 DEBUG SenderThread:30410 [sender.py:send():379] send: stats +2024-06-08 20:02:04,389 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: stop_status +2024-06-08 20:02:04,389 DEBUG SenderThread:30410 [sender.py:send_request():406] send_request: stop_status +2024-06-08 20:02:06,475 INFO Thread-12 :30410 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_190333-82mnef5m/files/output.log +2024-06-08 20:02:07,870 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 20:02:12,371 INFO Thread-12 :30410 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_190333-82mnef5m/files/output.log +2024-06-08 20:02:13,070 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 20:02:16,293 INFO Thread-12 :30410 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_190333-82mnef5m/files/output.log +2024-06-08 20:02:18,203 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 20:02:18,751 INFO Thread-12 :30410 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_190333-82mnef5m/files/output.log +2024-06-08 20:02:19,389 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: stop_status +2024-06-08 20:02:19,389 DEBUG SenderThread:30410 [sender.py:send_request():406] send_request: stop_status +2024-06-08 20:02:23,202 INFO Thread-12 :30410 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_190333-82mnef5m/files/output.log +2024-06-08 20:02:23,473 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 20:02:28,482 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 20:02:28,984 INFO Thread-12 :30410 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_190333-82mnef5m/files/output.log +2024-06-08 20:02:33,214 INFO Thread-12 :30410 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_190333-82mnef5m/files/output.log +2024-06-08 20:02:33,654 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 20:02:34,341 DEBUG SenderThread:30410 [sender.py:send():379] send: stats +2024-06-08 20:02:34,389 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: stop_status +2024-06-08 20:02:34,389 DEBUG SenderThread:30410 [sender.py:send_request():406] send_request: stop_status +2024-06-08 20:02:38,695 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 20:02:39,256 INFO Thread-12 :30410 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_190333-82mnef5m/files/output.log +2024-06-08 20:02:42,968 INFO Thread-12 :30410 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_190333-82mnef5m/files/output.log +2024-06-08 20:02:43,733 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 20:02:48,874 INFO Thread-12 :30410 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_190333-82mnef5m/files/output.log +2024-06-08 20:02:49,389 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: stop_status +2024-06-08 20:02:49,390 DEBUG SenderThread:30410 [sender.py:send_request():406] send_request: stop_status +2024-06-08 20:02:49,522 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 20:02:54,887 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 20:02:55,028 INFO Thread-12 :30410 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_190333-82mnef5m/files/output.log +2024-06-08 20:02:58,972 INFO Thread-12 :30410 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_190333-82mnef5m/files/output.log +2024-06-08 20:03:00,191 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 20:03:01,447 INFO Thread-12 :30410 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_190333-82mnef5m/files/output.log +2024-06-08 20:03:04,343 DEBUG SenderThread:30410 [sender.py:send():379] send: stats +2024-06-08 20:03:04,389 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: stop_status +2024-06-08 20:03:04,390 DEBUG SenderThread:30410 [sender.py:send_request():406] send_request: stop_status +2024-06-08 20:03:05,480 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 20:03:06,643 INFO Thread-12 :30410 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_190333-82mnef5m/files/output.log +2024-06-08 20:03:10,869 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 20:03:11,132 INFO Thread-12 :30410 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_190333-82mnef5m/files/output.log +2024-06-08 20:03:15,871 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 20:03:17,094 INFO Thread-12 :30410 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_190333-82mnef5m/files/output.log +2024-06-08 20:03:19,390 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: stop_status +2024-06-08 20:03:19,390 DEBUG SenderThread:30410 [sender.py:send_request():406] send_request: stop_status +2024-06-08 20:03:21,871 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 20:03:23,028 INFO Thread-12 :30410 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_190333-82mnef5m/files/output.log +2024-06-08 20:03:26,443 INFO Thread-12 :30410 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_190333-82mnef5m/files/output.log +2024-06-08 20:03:27,503 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 20:03:29,012 INFO Thread-12 :30410 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_190333-82mnef5m/files/output.log +2024-06-08 20:03:32,472 INFO Thread-12 :30410 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_190333-82mnef5m/files/output.log +2024-06-08 20:03:32,677 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 20:03:34,345 DEBUG SenderThread:30410 [sender.py:send():379] send: stats +2024-06-08 20:03:34,390 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: stop_status +2024-06-08 20:03:34,390 DEBUG SenderThread:30410 [sender.py:send_request():406] send_request: stop_status +2024-06-08 20:03:37,838 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 20:03:39,112 INFO Thread-12 :30410 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_190333-82mnef5m/files/output.log +2024-06-08 20:03:42,838 INFO Thread-12 :30410 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_190333-82mnef5m/files/output.log +2024-06-08 20:03:43,458 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 20:03:45,148 INFO Thread-12 :30410 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_190333-82mnef5m/files/output.log +2024-06-08 20:03:48,522 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 20:03:48,883 INFO Thread-12 :30410 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_190333-82mnef5m/files/output.log +2024-06-08 20:03:49,390 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: stop_status +2024-06-08 20:03:49,390 DEBUG SenderThread:30410 [sender.py:send_request():406] send_request: stop_status +2024-06-08 20:03:50,947 INFO Thread-12 :30410 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_190333-82mnef5m/files/output.log +2024-06-08 20:03:53,731 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 20:03:54,440 INFO Thread-12 :30410 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_190333-82mnef5m/files/output.log +2024-06-08 20:03:58,746 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 20:04:01,403 INFO Thread-12 :30410 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_190333-82mnef5m/files/output.log +2024-06-08 20:04:03,814 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 20:04:04,347 DEBUG SenderThread:30410 [sender.py:send():379] send: stats +2024-06-08 20:04:04,390 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: stop_status +2024-06-08 20:04:04,390 DEBUG SenderThread:30410 [sender.py:send_request():406] send_request: stop_status +2024-06-08 20:04:05,148 INFO Thread-12 :30410 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_190333-82mnef5m/files/output.log +2024-06-08 20:04:06,411 INFO Thread-12 :30410 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_190333-82mnef5m/files/output.log +2024-06-08 20:04:08,874 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 20:04:10,990 INFO Thread-12 :30410 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_190333-82mnef5m/files/output.log +2024-06-08 20:04:14,137 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 20:04:16,719 INFO Thread-12 :30410 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_190333-82mnef5m/files/output.log +2024-06-08 20:04:19,390 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: stop_status +2024-06-08 20:04:19,390 DEBUG SenderThread:30410 [sender.py:send_request():406] send_request: stop_status +2024-06-08 20:04:19,514 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 20:04:20,479 INFO Thread-12 :30410 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_190333-82mnef5m/files/output.log +2024-06-08 20:04:22,883 INFO Thread-12 :30410 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_190333-82mnef5m/files/output.log +2024-06-08 20:04:25,510 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 20:04:28,604 INFO Thread-12 :30410 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_190333-82mnef5m/files/output.log +2024-06-08 20:04:30,872 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 20:04:33,061 INFO Thread-12 :30410 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_190333-82mnef5m/files/output.log +2024-06-08 20:04:34,350 DEBUG SenderThread:30410 [sender.py:send():379] send: stats +2024-06-08 20:04:34,390 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: stop_status +2024-06-08 20:04:34,390 DEBUG SenderThread:30410 [sender.py:send_request():406] send_request: stop_status +2024-06-08 20:04:35,873 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 20:04:39,131 INFO Thread-12 :30410 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_190333-82mnef5m/files/output.log +2024-06-08 20:04:41,873 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 20:04:44,751 INFO Thread-12 :30410 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_190333-82mnef5m/files/output.log +2024-06-08 20:04:46,873 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 20:04:49,390 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: stop_status +2024-06-08 20:04:49,390 DEBUG SenderThread:30410 [sender.py:send_request():406] send_request: stop_status +2024-06-08 20:04:50,964 INFO Thread-12 :30410 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_190333-82mnef5m/files/output.log +2024-06-08 20:04:52,873 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 20:04:55,412 INFO Thread-12 :30410 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_190333-82mnef5m/files/output.log +2024-06-08 20:04:57,873 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 20:05:01,543 INFO Thread-12 :30410 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_190333-82mnef5m/files/output.log +2024-06-08 20:05:02,715 INFO Thread-12 :30410 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_190333-82mnef5m/files/output.log +2024-06-08 20:05:02,944 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 20:05:04,352 DEBUG SenderThread:30410 [sender.py:send():379] send: stats +2024-06-08 20:05:04,390 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: stop_status +2024-06-08 20:05:04,390 DEBUG SenderThread:30410 [sender.py:send_request():406] send_request: stop_status +2024-06-08 20:05:05,052 INFO Thread-12 :30410 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_190333-82mnef5m/files/output.log +2024-06-08 20:05:07,419 INFO Thread-12 :30410 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_190333-82mnef5m/files/output.log +2024-06-08 20:05:07,999 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 20:05:08,763 INFO Thread-12 :30410 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_190333-82mnef5m/files/output.log +2024-06-08 20:05:11,317 INFO Thread-12 :30410 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_190333-82mnef5m/files/output.log +2024-06-08 20:05:12,728 INFO Thread-12 :30410 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_190333-82mnef5m/files/output.log +2024-06-08 20:05:13,076 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 20:05:14,920 INFO Thread-12 :30410 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_190333-82mnef5m/files/output.log +2024-06-08 20:05:17,515 INFO Thread-12 :30410 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_190333-82mnef5m/files/output.log +2024-06-08 20:05:18,146 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 20:05:18,707 INFO Thread-12 :30410 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_190333-82mnef5m/files/output.log +2024-06-08 20:05:19,390 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: stop_status +2024-06-08 20:05:19,390 DEBUG SenderThread:30410 [sender.py:send_request():406] send_request: stop_status +2024-06-08 20:05:21,100 INFO Thread-12 :30410 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_190333-82mnef5m/files/output.log +2024-06-08 20:05:23,230 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 20:05:23,427 INFO Thread-12 :30410 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_190333-82mnef5m/files/output.log +2024-06-08 20:05:24,564 INFO Thread-12 :30410 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_190333-82mnef5m/files/output.log +2024-06-08 20:05:26,875 INFO Thread-12 :30410 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_190333-82mnef5m/files/output.log +2024-06-08 20:05:28,258 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 20:05:29,132 INFO Thread-12 :30410 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_190333-82mnef5m/files/output.log +2024-06-08 20:05:31,531 INFO Thread-12 :30410 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_190333-82mnef5m/files/output.log +2024-06-08 20:05:32,784 INFO Thread-12 :30410 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_190333-82mnef5m/files/output.log +2024-06-08 20:05:33,308 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 20:05:34,354 DEBUG SenderThread:30410 [sender.py:send():379] send: stats +2024-06-08 20:05:34,390 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: stop_status +2024-06-08 20:05:34,390 DEBUG SenderThread:30410 [sender.py:send_request():406] send_request: stop_status +2024-06-08 20:05:35,116 INFO Thread-12 :30410 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_190333-82mnef5m/files/output.log +2024-06-08 20:05:36,419 INFO Thread-12 :30410 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_190333-82mnef5m/files/output.log +2024-06-08 20:05:38,386 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 20:05:38,695 INFO Thread-12 :30410 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_190333-82mnef5m/files/output.log +2024-06-08 20:05:41,000 INFO Thread-12 :30410 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_190333-82mnef5m/files/output.log +2024-06-08 20:05:43,458 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 20:05:43,482 INFO Thread-12 :30410 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_190333-82mnef5m/files/output.log +2024-06-08 20:05:44,596 INFO Thread-12 :30410 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_190333-82mnef5m/files/output.log +2024-06-08 20:05:47,120 INFO Thread-12 :30410 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_190333-82mnef5m/files/output.log +2024-06-08 20:05:48,533 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 20:05:49,390 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: stop_status +2024-06-08 20:05:49,390 DEBUG SenderThread:30410 [sender.py:send_request():406] send_request: stop_status +2024-06-08 20:05:49,527 INFO Thread-12 :30410 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_190333-82mnef5m/files/output.log +2024-06-08 20:05:50,655 INFO Thread-12 :30410 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_190333-82mnef5m/files/output.log +2024-06-08 20:05:53,076 INFO Thread-12 :30410 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_190333-82mnef5m/files/output.log +2024-06-08 20:05:53,573 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 20:05:55,753 INFO Thread-12 :30410 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_190333-82mnef5m/files/output.log +2024-06-08 20:05:57,140 INFO Thread-12 :30410 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_190333-82mnef5m/files/output.log +2024-06-08 20:05:58,613 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 20:05:59,599 INFO Thread-12 :30410 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_190333-82mnef5m/files/output.log +2024-06-08 20:06:00,599 INFO Thread-12 :30410 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_190333-82mnef5m/files/output.log +2024-06-08 20:06:03,659 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 20:06:04,356 DEBUG SenderThread:30410 [sender.py:send():379] send: stats +2024-06-08 20:06:04,390 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: stop_status +2024-06-08 20:06:04,391 DEBUG SenderThread:30410 [sender.py:send_request():406] send_request: stop_status +2024-06-08 20:06:04,447 INFO Thread-12 :30410 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_190333-82mnef5m/files/output.log +2024-06-08 20:06:06,454 INFO Thread-12 :30410 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_190333-82mnef5m/files/output.log +2024-06-08 20:06:08,718 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 20:06:11,167 INFO Thread-12 :30410 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_190333-82mnef5m/files/output.log +2024-06-08 20:06:13,874 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 20:06:16,892 INFO Thread-12 :30410 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_190333-82mnef5m/files/output.log +2024-06-08 20:06:19,391 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: stop_status +2024-06-08 20:06:19,391 DEBUG SenderThread:30410 [sender.py:send_request():406] send_request: stop_status +2024-06-08 20:06:19,499 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 20:06:20,584 INFO Thread-12 :30410 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_190333-82mnef5m/files/output.log +2024-06-08 20:06:22,935 INFO Thread-12 :30410 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_190333-82mnef5m/files/output.log +2024-06-08 20:06:25,120 INFO Thread-12 :30410 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_190333-82mnef5m/files/output.log +2024-06-08 20:06:25,206 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 20:06:28,886 INFO Thread-12 :30410 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_190333-82mnef5m/files/output.log +2024-06-08 20:06:30,874 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 20:06:34,358 DEBUG SenderThread:30410 [sender.py:send():379] send: stats +2024-06-08 20:06:34,391 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: stop_status +2024-06-08 20:06:34,391 DEBUG SenderThread:30410 [sender.py:send_request():406] send_request: stop_status +2024-06-08 20:06:34,750 INFO Thread-12 :30410 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_190333-82mnef5m/files/output.log +2024-06-08 20:06:35,874 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 20:06:38,570 INFO Thread-12 :30410 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_190333-82mnef5m/files/output.log +2024-06-08 20:06:40,807 INFO Thread-12 :30410 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_190333-82mnef5m/files/output.log +2024-06-08 20:06:41,875 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 20:06:44,655 INFO Thread-12 :30410 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_190333-82mnef5m/files/output.log +2024-06-08 20:06:46,875 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 20:06:46,932 INFO Thread-12 :30410 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_190333-82mnef5m/files/output.log +2024-06-08 20:06:49,391 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: stop_status +2024-06-08 20:06:49,391 DEBUG SenderThread:30410 [sender.py:send_request():406] send_request: stop_status +2024-06-08 20:06:51,324 INFO Thread-12 :30410 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_190333-82mnef5m/files/output.log +2024-06-08 20:06:52,875 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 20:06:55,160 INFO Thread-12 :30410 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_190333-82mnef5m/files/output.log +2024-06-08 20:06:57,695 INFO Thread-12 :30410 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_190333-82mnef5m/files/output.log +2024-06-08 20:06:58,875 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 20:07:01,343 INFO Thread-12 :30410 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_190333-82mnef5m/files/output.log +2024-06-08 20:07:04,360 DEBUG SenderThread:30410 [sender.py:send():379] send: stats +2024-06-08 20:07:04,361 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 20:07:04,391 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: stop_status +2024-06-08 20:07:04,391 DEBUG SenderThread:30410 [sender.py:send_request():406] send_request: stop_status +2024-06-08 20:07:07,068 INFO Thread-12 :30410 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_190333-82mnef5m/files/output.log +2024-06-08 20:07:09,742 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 20:07:11,019 INFO Thread-12 :30410 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_190333-82mnef5m/files/output.log +2024-06-08 20:07:13,365 INFO Thread-12 :30410 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_190333-82mnef5m/files/output.log +2024-06-08 20:07:15,611 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 20:07:18,928 INFO Thread-12 :30410 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_190333-82mnef5m/files/output.log +2024-06-08 20:07:19,391 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: stop_status +2024-06-08 20:07:19,391 DEBUG SenderThread:30410 [sender.py:send_request():406] send_request: stop_status +2024-06-08 20:07:20,875 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 20:07:23,711 INFO Thread-12 :30410 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_190333-82mnef5m/files/output.log +2024-06-08 20:07:25,875 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 20:07:29,716 INFO Thread-12 :30410 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_190333-82mnef5m/files/output.log +2024-06-08 20:07:30,876 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 20:07:33,334 INFO Thread-12 :30410 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_190333-82mnef5m/files/output.log +2024-06-08 20:07:34,362 DEBUG SenderThread:30410 [sender.py:send():379] send: stats +2024-06-08 20:07:34,391 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: stop_status +2024-06-08 20:07:34,391 DEBUG SenderThread:30410 [sender.py:send_request():406] send_request: stop_status +2024-06-08 20:07:35,876 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 20:07:39,389 INFO Thread-12 :30410 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_190333-82mnef5m/files/output.log +2024-06-08 20:07:40,876 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 20:07:42,735 INFO Thread-12 :30410 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_190333-82mnef5m/files/output.log +2024-06-08 20:07:44,831 INFO Thread-12 :30410 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_190333-82mnef5m/files/output.log +2024-06-08 20:07:45,876 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 20:07:49,391 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: stop_status +2024-06-08 20:07:49,391 DEBUG SenderThread:30410 [sender.py:send_request():406] send_request: stop_status +2024-06-08 20:07:49,803 INFO Thread-12 :30410 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_190333-82mnef5m/files/output.log +2024-06-08 20:07:50,876 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 20:07:55,559 INFO Thread-12 :30410 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_190333-82mnef5m/files/output.log +2024-06-08 20:07:55,876 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 20:07:59,040 INFO Thread-12 :30410 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_190333-82mnef5m/files/output.log +2024-06-08 20:08:00,876 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 20:08:01,112 INFO Thread-12 :30410 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_190333-82mnef5m/files/output.log +2024-06-08 20:08:04,364 DEBUG SenderThread:30410 [sender.py:send():379] send: stats +2024-06-08 20:08:04,391 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: stop_status +2024-06-08 20:08:04,391 DEBUG SenderThread:30410 [sender.py:send_request():406] send_request: stop_status +2024-06-08 20:08:04,903 INFO Thread-12 :30410 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_190333-82mnef5m/files/output.log +2024-06-08 20:08:05,876 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 20:08:10,876 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 20:08:11,024 INFO Thread-12 :30410 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_190333-82mnef5m/files/output.log +2024-06-08 20:08:14,855 INFO Thread-12 :30410 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_190333-82mnef5m/files/output.log +2024-06-08 20:08:15,876 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 20:08:19,391 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: stop_status +2024-06-08 20:08:19,392 DEBUG SenderThread:30410 [sender.py:send_request():406] send_request: stop_status +2024-06-08 20:08:20,877 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 20:08:21,130 INFO Thread-12 :30410 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_190333-82mnef5m/files/output.log +2024-06-08 20:08:24,835 INFO Thread-12 :30410 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_190333-82mnef5m/files/output.log +2024-06-08 20:08:25,877 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 20:08:30,748 INFO Thread-12 :30410 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_190333-82mnef5m/files/output.log +2024-06-08 20:08:30,877 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 20:08:34,366 DEBUG SenderThread:30410 [sender.py:send():379] send: stats +2024-06-08 20:08:34,391 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: stop_status +2024-06-08 20:08:34,391 DEBUG SenderThread:30410 [sender.py:send_request():406] send_request: stop_status +2024-06-08 20:08:35,479 INFO Thread-12 :30410 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_190333-82mnef5m/files/output.log +2024-06-08 20:08:35,877 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 20:08:36,835 INFO Thread-12 :30410 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_190333-82mnef5m/files/output.log +2024-06-08 20:08:40,877 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 20:08:41,625 INFO Thread-12 :30410 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_190333-82mnef5m/files/output.log +2024-06-08 20:08:45,882 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 20:08:46,719 INFO Thread-12 :30410 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_190333-82mnef5m/files/output.log +2024-06-08 20:08:49,391 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: stop_status +2024-06-08 20:08:49,392 DEBUG SenderThread:30410 [sender.py:send_request():406] send_request: stop_status +2024-06-08 20:08:50,959 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 20:08:51,542 INFO Thread-12 :30410 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_190333-82mnef5m/files/output.log +2024-06-08 20:08:52,694 INFO Thread-12 :30410 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_190333-82mnef5m/files/output.log +2024-06-08 20:08:56,121 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 20:08:57,595 INFO Thread-12 :30410 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_190333-82mnef5m/files/output.log +2024-06-08 20:08:58,932 INFO Thread-12 :30410 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_190333-82mnef5m/files/output.log +2024-06-08 20:09:01,527 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 20:09:02,671 INFO Thread-12 :30410 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_190333-82mnef5m/files/output.log +2024-06-08 20:09:04,368 DEBUG SenderThread:30410 [sender.py:send():379] send: stats +2024-06-08 20:09:04,392 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: stop_status +2024-06-08 20:09:04,392 DEBUG SenderThread:30410 [sender.py:send_request():406] send_request: stop_status +2024-06-08 20:09:05,008 INFO Thread-12 :30410 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_190333-82mnef5m/files/output.log +2024-06-08 20:09:06,877 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 20:09:09,341 INFO Thread-12 :30410 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_190333-82mnef5m/files/output.log +2024-06-08 20:09:11,878 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 20:09:15,313 INFO Thread-12 :30410 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_190333-82mnef5m/files/output.log +2024-06-08 20:09:17,357 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 20:09:18,831 INFO Thread-12 :30410 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_190333-82mnef5m/files/output.log +2024-06-08 20:09:19,392 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: stop_status +2024-06-08 20:09:19,392 DEBUG SenderThread:30410 [sender.py:send_request():406] send_request: stop_status +2024-06-08 20:09:22,391 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 20:09:25,044 INFO Thread-12 :30410 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_190333-82mnef5m/files/output.log +2024-06-08 20:09:27,414 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 20:09:28,950 INFO Thread-12 :30410 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_190333-82mnef5m/files/output.log +2024-06-08 20:09:31,571 INFO Thread-12 :30410 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_190333-82mnef5m/files/output.log +2024-06-08 20:09:32,878 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 20:09:34,370 DEBUG SenderThread:30410 [sender.py:send():379] send: stats +2024-06-08 20:09:34,392 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: stop_status +2024-06-08 20:09:34,392 DEBUG SenderThread:30410 [sender.py:send_request():406] send_request: stop_status +2024-06-08 20:09:35,446 INFO Thread-12 :30410 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_190333-82mnef5m/files/output.log +2024-06-08 20:09:36,827 INFO Thread-12 :30410 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_190333-82mnef5m/files/output.log +2024-06-08 20:09:38,878 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 20:09:41,641 INFO Thread-12 :30410 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_190333-82mnef5m/files/output.log +2024-06-08 20:09:43,924 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 20:09:45,209 INFO Thread-12 :30410 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_190333-82mnef5m/files/output.log +2024-06-08 20:09:48,963 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 20:09:49,392 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: stop_status +2024-06-08 20:09:49,392 DEBUG SenderThread:30410 [sender.py:send_request():406] send_request: stop_status +2024-06-08 20:09:51,161 INFO Thread-12 :30410 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_190333-82mnef5m/files/output.log +2024-06-08 20:09:54,780 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 20:09:55,696 INFO Thread-12 :30410 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_190333-82mnef5m/files/output.log +2024-06-08 20:10:00,546 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 20:10:00,695 INFO Thread-12 :30410 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_190333-82mnef5m/files/output.log +2024-06-08 20:10:04,373 DEBUG SenderThread:30410 [sender.py:send():379] send: stats +2024-06-08 20:10:04,392 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: stop_status +2024-06-08 20:10:04,392 DEBUG SenderThread:30410 [sender.py:send_request():406] send_request: stop_status +2024-06-08 20:10:05,188 INFO Thread-12 :30410 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_190333-82mnef5m/files/output.log +2024-06-08 20:10:05,559 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 20:10:07,443 INFO Thread-12 :30410 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_190333-82mnef5m/files/output.log +2024-06-08 20:10:10,883 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 20:10:10,948 INFO Thread-12 :30410 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_190333-82mnef5m/files/output.log +2024-06-08 20:10:15,503 INFO Thread-12 :30410 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_190333-82mnef5m/files/output.log +2024-06-08 20:10:16,882 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 20:10:19,392 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: stop_status +2024-06-08 20:10:19,392 DEBUG SenderThread:30410 [sender.py:send_request():406] send_request: stop_status +2024-06-08 20:10:21,767 INFO Thread-12 :30410 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_190333-82mnef5m/files/output.log +2024-06-08 20:10:22,883 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 20:10:25,274 INFO Thread-12 :30410 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_190333-82mnef5m/files/output.log +2024-06-08 20:10:27,859 INFO Thread-12 :30410 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_190333-82mnef5m/files/output.log +2024-06-08 20:10:28,883 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 20:10:33,339 INFO Thread-12 :30410 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_190333-82mnef5m/files/output.log +2024-06-08 20:10:34,374 DEBUG SenderThread:30410 [sender.py:send():379] send: stats +2024-06-08 20:10:34,375 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 20:10:34,392 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: stop_status +2024-06-08 20:10:34,392 DEBUG SenderThread:30410 [sender.py:send_request():406] send_request: stop_status +2024-06-08 20:10:36,919 INFO Thread-12 :30410 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_190333-82mnef5m/files/output.log +2024-06-08 20:10:39,883 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 20:10:43,669 INFO Thread-12 :30410 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_190333-82mnef5m/files/output.log +2024-06-08 20:10:44,911 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 20:10:46,972 INFO Thread-12 :30410 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_190333-82mnef5m/files/output.log +2024-06-08 20:10:49,136 INFO Thread-12 :30410 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_190333-82mnef5m/files/output.log +2024-06-08 20:10:49,396 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: stop_status +2024-06-08 20:10:49,396 DEBUG SenderThread:30410 [sender.py:send_request():406] send_request: stop_status +2024-06-08 20:10:50,453 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 20:10:53,595 INFO Thread-12 :30410 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_190333-82mnef5m/files/output.log +2024-06-08 20:10:55,456 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 20:10:57,020 INFO Thread-12 :30410 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_190333-82mnef5m/files/output.log +2024-06-08 20:10:59,352 INFO Thread-12 :30410 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_190333-82mnef5m/files/output.log +2024-06-08 20:11:00,799 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 20:11:03,059 INFO Thread-12 :30410 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_190333-82mnef5m/files/output.log +2024-06-08 20:11:04,376 DEBUG SenderThread:30410 [sender.py:send():379] send: stats +2024-06-08 20:11:04,392 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: stop_status +2024-06-08 20:11:04,392 DEBUG SenderThread:30410 [sender.py:send_request():406] send_request: stop_status +2024-06-08 20:11:05,883 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 20:11:08,924 INFO Thread-12 :30410 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_190333-82mnef5m/files/output.log +2024-06-08 20:11:10,884 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 20:11:13,775 INFO Thread-12 :30410 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_190333-82mnef5m/files/output.log +2024-06-08 20:11:15,884 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 20:11:17,743 INFO Thread-12 :30410 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_190333-82mnef5m/files/output.log +2024-06-08 20:11:19,012 INFO Thread-12 :30410 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_190333-82mnef5m/files/output.log +2024-06-08 20:11:19,392 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: stop_status +2024-06-08 20:11:19,393 DEBUG SenderThread:30410 [sender.py:send_request():406] send_request: stop_status +2024-06-08 20:11:20,883 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 20:11:23,270 INFO Thread-12 :30410 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_190333-82mnef5m/files/output.log +2024-06-08 20:11:25,884 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 20:11:29,340 INFO Thread-12 :30410 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_190333-82mnef5m/files/output.log +2024-06-08 20:11:30,884 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 20:11:33,003 INFO Thread-12 :30410 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_190333-82mnef5m/files/output.log +2024-06-08 20:11:34,379 DEBUG SenderThread:30410 [sender.py:send():379] send: stats +2024-06-08 20:11:34,392 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: stop_status +2024-06-08 20:11:34,393 DEBUG SenderThread:30410 [sender.py:send_request():406] send_request: stop_status +2024-06-08 20:11:35,884 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 20:11:37,853 INFO Thread-12 :30410 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_190333-82mnef5m/files/output.log +2024-06-08 20:11:39,040 INFO Thread-12 :30410 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_190333-82mnef5m/files/output.log +2024-06-08 20:11:40,884 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 20:11:43,831 INFO Thread-12 :30410 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_190333-82mnef5m/files/output.log +2024-06-08 20:11:45,884 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 20:11:47,225 INFO Thread-12 :30410 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_190333-82mnef5m/files/output.log +2024-06-08 20:11:49,307 INFO Thread-12 :30410 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_190333-82mnef5m/files/output.log +2024-06-08 20:11:49,392 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: stop_status +2024-06-08 20:11:49,393 DEBUG SenderThread:30410 [sender.py:send_request():406] send_request: stop_status +2024-06-08 20:11:50,884 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 20:11:53,155 INFO Thread-12 :30410 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_190333-82mnef5m/files/output.log +2024-06-08 20:11:56,424 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 20:11:56,883 INFO Thread-12 :30410 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_190333-82mnef5m/files/output.log +2024-06-08 20:11:59,245 INFO Thread-12 :30410 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_190333-82mnef5m/files/output.log +2024-06-08 20:12:01,516 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 20:12:03,580 INFO Thread-12 :30410 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_190333-82mnef5m/files/output.log +2024-06-08 20:12:04,381 DEBUG SenderThread:30410 [sender.py:send():379] send: stats +2024-06-08 20:12:04,393 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: stop_status +2024-06-08 20:12:04,393 DEBUG SenderThread:30410 [sender.py:send_request():406] send_request: stop_status +2024-06-08 20:12:06,548 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 20:12:09,575 INFO Thread-12 :30410 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_190333-82mnef5m/files/output.log +2024-06-08 20:12:11,682 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 20:12:13,184 INFO Thread-12 :30410 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_190333-82mnef5m/files/output.log +2024-06-08 20:12:16,689 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 20:12:19,124 INFO Thread-12 :30410 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_190333-82mnef5m/files/output.log +2024-06-08 20:12:19,393 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: stop_status +2024-06-08 20:12:19,393 DEBUG SenderThread:30410 [sender.py:send_request():406] send_request: stop_status +2024-06-08 20:12:22,629 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 20:12:22,919 INFO Thread-12 :30410 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_190333-82mnef5m/files/output.log +2024-06-08 20:12:27,651 INFO Thread-12 :30410 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_190333-82mnef5m/files/output.log +2024-06-08 20:12:28,559 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 20:12:33,233 INFO Thread-12 :30410 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_190333-82mnef5m/files/output.log +2024-06-08 20:12:34,124 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 20:12:34,383 DEBUG SenderThread:30410 [sender.py:send():379] send: stats +2024-06-08 20:12:34,393 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: stop_status +2024-06-08 20:12:34,393 DEBUG SenderThread:30410 [sender.py:send_request():406] send_request: stop_status +2024-06-08 20:12:37,775 INFO Thread-12 :30410 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_190333-82mnef5m/files/output.log +2024-06-08 20:12:39,876 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 20:12:41,439 INFO Thread-12 :30410 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_190333-82mnef5m/files/output.log +2024-06-08 20:12:45,028 INFO Thread-12 :30410 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_190333-82mnef5m/files/output.log +2024-06-08 20:12:45,734 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 20:12:47,316 INFO Thread-12 :30410 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_190333-82mnef5m/files/output.log +2024-06-08 20:12:49,393 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: stop_status +2024-06-08 20:12:49,393 DEBUG SenderThread:30410 [sender.py:send_request():406] send_request: stop_status +2024-06-08 20:12:50,885 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 20:12:52,039 INFO Thread-12 :30410 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_190333-82mnef5m/files/output.log +2024-06-08 20:12:55,827 INFO Thread-12 :30410 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_190333-82mnef5m/files/output.log +2024-06-08 20:12:55,885 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 20:13:00,886 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 20:13:04,385 DEBUG SenderThread:30410 [sender.py:send():379] send: stats +2024-06-08 20:13:04,423 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: stop_status +2024-06-08 20:13:04,424 DEBUG SenderThread:30410 [sender.py:send_request():406] send_request: stop_status +2024-06-08 20:13:06,136 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 20:13:12,098 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 20:13:17,116 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 20:13:19,547 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: stop_status +2024-06-08 20:13:19,547 DEBUG SenderThread:30410 [sender.py:send_request():406] send_request: stop_status +2024-06-08 20:13:22,164 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 20:13:27,184 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 20:13:31,088 INFO Thread-12 :30410 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_190333-82mnef5m/files/output.log +2024-06-08 20:13:32,131 INFO Thread-12 :30410 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_190333-82mnef5m/files/output.log +2024-06-08 20:13:32,315 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 20:13:34,387 DEBUG SenderThread:30410 [sender.py:send():379] send: stats +2024-06-08 20:13:34,546 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: stop_status +2024-06-08 20:13:34,547 DEBUG SenderThread:30410 [sender.py:send_request():406] send_request: stop_status +2024-06-08 20:13:38,315 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 20:13:44,316 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 20:13:49,546 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: stop_status +2024-06-08 20:13:49,547 DEBUG SenderThread:30410 [sender.py:send_request():406] send_request: stop_status +2024-06-08 20:13:49,614 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 20:13:55,316 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 20:14:01,316 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 20:14:04,389 DEBUG SenderThread:30410 [sender.py:send():379] send: stats +2024-06-08 20:14:04,546 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: stop_status +2024-06-08 20:14:04,547 DEBUG SenderThread:30410 [sender.py:send_request():406] send_request: stop_status +2024-06-08 20:14:06,316 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 20:14:12,316 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 20:14:18,316 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 20:14:19,547 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: stop_status +2024-06-08 20:14:19,547 DEBUG SenderThread:30410 [sender.py:send_request():406] send_request: stop_status +2024-06-08 20:14:24,316 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 20:14:30,317 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 20:14:34,392 DEBUG SenderThread:30410 [sender.py:send():379] send: stats +2024-06-08 20:14:34,547 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: stop_status +2024-06-08 20:14:34,547 DEBUG SenderThread:30410 [sender.py:send_request():406] send_request: stop_status +2024-06-08 20:14:35,712 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 20:14:41,317 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 20:14:46,317 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 20:14:49,547 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: stop_status +2024-06-08 20:14:49,547 DEBUG SenderThread:30410 [sender.py:send_request():406] send_request: stop_status +2024-06-08 20:14:51,317 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 20:14:56,317 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 20:15:01,317 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 20:15:04,394 DEBUG SenderThread:30410 [sender.py:send():379] send: stats +2024-06-08 20:15:04,547 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: stop_status +2024-06-08 20:15:04,547 DEBUG SenderThread:30410 [sender.py:send_request():406] send_request: stop_status +2024-06-08 20:15:06,317 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 20:15:11,317 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 20:15:16,317 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 20:15:19,547 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: stop_status +2024-06-08 20:15:19,547 DEBUG SenderThread:30410 [sender.py:send_request():406] send_request: stop_status +2024-06-08 20:15:21,317 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 20:15:26,317 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 20:15:31,318 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 20:15:34,396 DEBUG SenderThread:30410 [sender.py:send():379] send: stats +2024-06-08 20:15:34,547 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: stop_status +2024-06-08 20:15:34,547 DEBUG SenderThread:30410 [sender.py:send_request():406] send_request: stop_status +2024-06-08 20:15:36,318 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 20:15:39,081 INFO Thread-12 :30410 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_190333-82mnef5m/files/output.log +2024-06-08 20:15:41,384 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 20:15:41,599 INFO Thread-12 :30410 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_190333-82mnef5m/files/output.log +2024-06-08 20:15:46,384 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 20:15:49,547 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: stop_status +2024-06-08 20:15:49,548 DEBUG SenderThread:30410 [sender.py:send_request():406] send_request: stop_status +2024-06-08 20:15:51,243 DEBUG SenderThread:30410 [sender.py:send():379] send: config +2024-06-08 20:15:51,244 DEBUG SenderThread:30410 [sender.py:send():379] send: telemetry +2024-06-08 20:15:51,245 DEBUG SenderThread:30410 [sender.py:send_request():406] send_request: summary_record +2024-06-08 20:15:51,246 INFO SenderThread:30410 [sender.py:_save_file():1454] saving file wandb-summary.json with policy end +2024-06-08 20:15:51,250 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: partial_history +2024-06-08 20:15:51,253 DEBUG SenderThread:30410 [sender.py:send():379] send: history +2024-06-08 20:15:51,254 DEBUG SenderThread:30410 [sender.py:send_request():406] send_request: summary_record +2024-06-08 20:15:51,256 INFO SenderThread:30410 [sender.py:_save_file():1454] saving file wandb-summary.json with policy end +2024-06-08 20:15:51,484 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: log_artifact +2024-06-08 20:15:51,484 DEBUG SenderThread:30410 [sender.py:send_request():406] send_request: log_artifact +2024-06-08 20:15:51,623 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: partial_history +2024-06-08 20:15:51,712 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: log_artifact +2024-06-08 20:15:51,836 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: server_info +2024-06-08 20:15:51,837 INFO Thread-12 :30410 [dir_watcher.py:_on_file_created():271] file/dir created: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_190333-82mnef5m/files/wandb-summary.json +2024-06-08 20:15:51,837 INFO Thread-12 :30410 [dir_watcher.py:_on_file_created():271] file/dir created: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_190333-82mnef5m/files/media/table/evaluation/eval_results_1_fd1718bec4834f9c9150.table.json +2024-06-08 20:15:51,838 INFO Thread-12 :30410 [dir_watcher.py:_on_file_created():271] file/dir created: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_190333-82mnef5m/files/media/table +2024-06-08 20:15:51,838 INFO Thread-12 :30410 [dir_watcher.py:_on_file_created():271] file/dir created: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_190333-82mnef5m/files/media +2024-06-08 20:15:51,838 INFO Thread-12 :30410 [dir_watcher.py:_on_file_created():271] file/dir created: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_190333-82mnef5m/files/media/table/evaluation +2024-06-08 20:15:52,240 INFO wandb-upload_0:30410 [upload_job.py:push():88] Uploaded file /root/.local/share/wandb/artifacts/staging/tmph7fs69sz +2024-06-08 20:15:52,280 INFO wandb-upload_1:30410 [upload_job.py:push():130] Uploaded file /tmp/tmpggzsy1awwandb/i15ssb18-media/table/evaluation/eval_results_1_fd1718bec4834f9c9150.table.json +2024-06-08 20:15:52,685 INFO SenderThread:30410 [sender.py:send_request_log_artifact():1518] logged artifact run-82mnef5m-evaluationeval_results - {'id': 'QXJ0aWZhY3Q6ODY1MDg0NzE4', 'state': 'PENDING', 'artifactSequence': {'id': 'QXJ0aWZhY3RDb2xsZWN0aW9uOjE4MzczOTM1Nw==', 'latestArtifact': None}} +2024-06-08 20:15:52,685 DEBUG SenderThread:30410 [sender.py:send():379] send: files +2024-06-08 20:15:52,685 INFO SenderThread:30410 [sender.py:_save_file():1454] saving file media/table/evaluation/eval_results_1_fd1718bec4834f9c9150.table.json with policy now +2024-06-08 20:15:52,685 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 20:15:52,686 DEBUG SenderThread:30410 [sender.py:send():379] send: history +2024-06-08 20:15:52,686 DEBUG SenderThread:30410 [sender.py:send_request():406] send_request: summary_record +2024-06-08 20:15:52,689 INFO SenderThread:30410 [sender.py:_save_file():1454] saving file wandb-summary.json with policy end +2024-06-08 20:15:52,689 DEBUG SenderThread:30410 [sender.py:send_request():406] send_request: log_artifact +2024-06-08 20:15:52,839 INFO Thread-12 :30410 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_190333-82mnef5m/files/wandb-summary.json +2024-06-08 20:15:53,007 INFO wandb-upload_0:30410 [upload_job.py:push():130] Uploaded file /tmp/tmpggzsy1awwandb/bpc4r29p-media/table/evaluation/eval_results_1_fd1718bec4834f9c9150.table.json +2024-06-08 20:15:53,385 INFO wandb-upload_1:30410 [upload_job.py:push():88] Uploaded file /tmp/tmpn16dc37k/results.json +2024-06-08 20:15:53,853 INFO SenderThread:30410 [sender.py:send_request_log_artifact():1518] logged artifact results - {'id': 'QXJ0aWZhY3Q6ODY1MDg0NzIz', 'state': 'PENDING', 'artifactSequence': {'id': 'QXJ0aWZhY3RDb2xsZWN0aW9uOjE4MjQwOTE0OQ==', 'latestArtifact': {'id': 'QXJ0aWZhY3Q6ODY1MDM2ODMz', 'versionIndex': 14}}} +2024-06-08 20:15:53,853 DEBUG SenderThread:30410 [sender.py:send():379] send: telemetry +2024-06-08 20:15:53,854 DEBUG SenderThread:30410 [sender.py:send_request():406] send_request: server_info +2024-06-08 20:15:53,906 DEBUG SenderThread:30410 [sender.py:send():379] send: exit +2024-06-08 20:15:53,906 INFO SenderThread:30410 [sender.py:send_exit():586] handling exit code: 0 +2024-06-08 20:15:53,906 INFO SenderThread:30410 [sender.py:send_exit():588] handling runtime: 4339 +2024-06-08 20:15:53,907 INFO SenderThread:30410 [sender.py:_save_file():1454] saving file wandb-summary.json with policy end +2024-06-08 20:15:53,907 INFO SenderThread:30410 [sender.py:send_exit():594] send defer +2024-06-08 20:15:53,908 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: defer +2024-06-08 20:15:53,908 INFO HandlerThread:30410 [handler.py:handle_request_defer():184] handle defer: 0 +2024-06-08 20:15:53,908 DEBUG SenderThread:30410 [sender.py:send_request():406] send_request: defer +2024-06-08 20:15:53,908 INFO SenderThread:30410 [sender.py:send_request_defer():610] handle sender defer: 0 +2024-06-08 20:15:53,908 INFO SenderThread:30410 [sender.py:transition_state():614] send defer: 1 +2024-06-08 20:15:53,908 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: defer +2024-06-08 20:15:53,908 INFO HandlerThread:30410 [handler.py:handle_request_defer():184] handle defer: 1 +2024-06-08 20:15:53,908 DEBUG SenderThread:30410 [sender.py:send_request():406] send_request: defer +2024-06-08 20:15:53,908 INFO SenderThread:30410 [sender.py:send_request_defer():610] handle sender defer: 1 +2024-06-08 20:15:53,908 INFO SenderThread:30410 [sender.py:transition_state():614] send defer: 2 +2024-06-08 20:15:53,908 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: defer +2024-06-08 20:15:53,908 INFO HandlerThread:30410 [handler.py:handle_request_defer():184] handle defer: 2 +2024-06-08 20:15:53,908 INFO HandlerThread:30410 [system_monitor.py:finish():203] Stopping system monitor +2024-06-08 20:15:53,908 DEBUG SystemMonitor:30410 [system_monitor.py:_start():179] Finished system metrics aggregation loop +2024-06-08 20:15:53,908 DEBUG SystemMonitor:30410 [system_monitor.py:_start():183] Publishing last batch of metrics +2024-06-08 20:15:53,911 INFO HandlerThread:30410 [interfaces.py:finish():200] Joined cpu monitor +2024-06-08 20:15:53,911 INFO HandlerThread:30410 [interfaces.py:finish():200] Joined disk monitor +2024-06-08 20:15:53,911 INFO HandlerThread:30410 [interfaces.py:finish():200] Joined memory monitor +2024-06-08 20:15:53,911 INFO HandlerThread:30410 [interfaces.py:finish():200] Joined network monitor +2024-06-08 20:15:53,912 DEBUG SenderThread:30410 [sender.py:send_request():406] send_request: defer +2024-06-08 20:15:53,912 INFO SenderThread:30410 [sender.py:send_request_defer():610] handle sender defer: 2 +2024-06-08 20:15:53,912 INFO SenderThread:30410 [sender.py:transition_state():614] send defer: 3 +2024-06-08 20:15:53,912 DEBUG SenderThread:30410 [sender.py:send():379] send: stats +2024-06-08 20:15:53,912 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: defer +2024-06-08 20:15:53,913 INFO HandlerThread:30410 [handler.py:handle_request_defer():184] handle defer: 3 +2024-06-08 20:15:53,913 DEBUG SenderThread:30410 [sender.py:send_request():406] send_request: defer +2024-06-08 20:15:53,913 INFO SenderThread:30410 [sender.py:send_request_defer():610] handle sender defer: 3 +2024-06-08 20:15:53,913 INFO SenderThread:30410 [sender.py:transition_state():614] send defer: 4 +2024-06-08 20:15:53,913 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: defer +2024-06-08 20:15:53,913 INFO HandlerThread:30410 [handler.py:handle_request_defer():184] handle defer: 4 +2024-06-08 20:15:53,913 DEBUG SenderThread:30410 [sender.py:send_request():406] send_request: defer +2024-06-08 20:15:53,913 INFO SenderThread:30410 [sender.py:send_request_defer():610] handle sender defer: 4 +2024-06-08 20:15:53,913 INFO SenderThread:30410 [sender.py:transition_state():614] send defer: 5 +2024-06-08 20:15:53,913 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: defer +2024-06-08 20:15:53,913 INFO HandlerThread:30410 [handler.py:handle_request_defer():184] handle defer: 5 +2024-06-08 20:15:53,914 DEBUG SenderThread:30410 [sender.py:send():379] send: summary +2024-06-08 20:15:53,915 INFO SenderThread:30410 [sender.py:_save_file():1454] saving file wandb-summary.json with policy end +2024-06-08 20:15:53,915 DEBUG SenderThread:30410 [sender.py:send_request():406] send_request: defer +2024-06-08 20:15:53,915 INFO SenderThread:30410 [sender.py:send_request_defer():610] handle sender defer: 5 +2024-06-08 20:15:53,915 INFO SenderThread:30410 [sender.py:transition_state():614] send defer: 6 +2024-06-08 20:15:53,915 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: defer +2024-06-08 20:15:53,915 INFO HandlerThread:30410 [handler.py:handle_request_defer():184] handle defer: 6 +2024-06-08 20:15:53,916 DEBUG SenderThread:30410 [sender.py:send_request():406] send_request: defer +2024-06-08 20:15:53,916 INFO SenderThread:30410 [sender.py:send_request_defer():610] handle sender defer: 6 +2024-06-08 20:15:53,920 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 20:15:54,139 INFO SenderThread:30410 [sender.py:transition_state():614] send defer: 7 +2024-06-08 20:15:54,139 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: defer +2024-06-08 20:15:54,139 INFO HandlerThread:30410 [handler.py:handle_request_defer():184] handle defer: 7 +2024-06-08 20:15:54,139 DEBUG SenderThread:30410 [sender.py:send_request():406] send_request: defer +2024-06-08 20:15:54,139 INFO SenderThread:30410 [sender.py:send_request_defer():610] handle sender defer: 7 +2024-06-08 20:15:54,843 INFO Thread-12 :30410 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_190333-82mnef5m/files/wandb-summary.json +2024-06-08 20:15:54,843 INFO Thread-12 :30410 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_190333-82mnef5m/files/config.yaml +2024-06-08 20:15:54,906 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: poll_exit +2024-06-08 20:15:55,388 INFO SenderThread:30410 [sender.py:transition_state():614] send defer: 8 +2024-06-08 20:15:55,388 DEBUG SenderThread:30410 [sender.py:send_request():406] send_request: poll_exit +2024-06-08 20:15:55,388 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: defer +2024-06-08 20:15:55,389 INFO HandlerThread:30410 [handler.py:handle_request_defer():184] handle defer: 8 +2024-06-08 20:15:55,389 DEBUG SenderThread:30410 [sender.py:send_request():406] send_request: defer +2024-06-08 20:15:55,389 INFO SenderThread:30410 [sender.py:send_request_defer():610] handle sender defer: 8 +2024-06-08 20:15:55,389 INFO SenderThread:30410 [job_builder.py:build():440] Attempting to build job artifact +2024-06-08 20:15:55,391 INFO SenderThread:30410 [job_builder.py:_get_source_type():580] no source found +2024-06-08 20:15:55,391 INFO SenderThread:30410 [sender.py:transition_state():614] send defer: 9 +2024-06-08 20:15:55,392 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: defer +2024-06-08 20:15:55,392 INFO HandlerThread:30410 [handler.py:handle_request_defer():184] handle defer: 9 +2024-06-08 20:15:55,392 DEBUG SenderThread:30410 [sender.py:send_request():406] send_request: defer +2024-06-08 20:15:55,392 INFO SenderThread:30410 [sender.py:send_request_defer():610] handle sender defer: 9 +2024-06-08 20:15:55,392 INFO SenderThread:30410 [dir_watcher.py:finish():358] shutting down directory watcher +2024-06-08 20:15:55,845 INFO SenderThread:30410 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_190333-82mnef5m/files/output.log +2024-06-08 20:15:55,846 INFO SenderThread:30410 [dir_watcher.py:finish():388] scan: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_190333-82mnef5m/files +2024-06-08 20:15:55,846 INFO SenderThread:30410 [dir_watcher.py:finish():402] scan save: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_190333-82mnef5m/files/wandb-metadata.json wandb-metadata.json +2024-06-08 20:15:55,846 INFO SenderThread:30410 [dir_watcher.py:finish():402] scan save: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_190333-82mnef5m/files/config.yaml config.yaml +2024-06-08 20:15:55,846 INFO SenderThread:30410 [dir_watcher.py:finish():402] scan save: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_190333-82mnef5m/files/output.log output.log +2024-06-08 20:15:55,848 INFO SenderThread:30410 [dir_watcher.py:finish():402] scan save: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_190333-82mnef5m/files/requirements.txt requirements.txt +2024-06-08 20:15:55,850 INFO SenderThread:30410 [dir_watcher.py:finish():402] scan save: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_190333-82mnef5m/files/wandb-summary.json wandb-summary.json +2024-06-08 20:15:55,851 INFO SenderThread:30410 [dir_watcher.py:finish():402] scan save: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_190333-82mnef5m/files/media/table/evaluation/eval_results_1_fd1718bec4834f9c9150.table.json media/table/evaluation/eval_results_1_fd1718bec4834f9c9150.table.json +2024-06-08 20:15:55,851 INFO SenderThread:30410 [sender.py:transition_state():614] send defer: 10 +2024-06-08 20:15:55,851 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: defer +2024-06-08 20:15:55,851 INFO HandlerThread:30410 [handler.py:handle_request_defer():184] handle defer: 10 +2024-06-08 20:15:55,851 DEBUG SenderThread:30410 [sender.py:send_request():406] send_request: defer +2024-06-08 20:15:55,851 INFO SenderThread:30410 [sender.py:send_request_defer():610] handle sender defer: 10 +2024-06-08 20:15:55,851 INFO SenderThread:30410 [file_pusher.py:finish():169] shutting down file pusher +2024-06-08 20:15:55,907 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: poll_exit +2024-06-08 20:15:55,907 DEBUG SenderThread:30410 [sender.py:send_request():406] send_request: poll_exit +2024-06-08 20:15:56,064 INFO wandb-upload_0:30410 [upload_job.py:push():130] Uploaded file /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_190333-82mnef5m/files/config.yaml +2024-06-08 20:15:56,078 INFO wandb-upload_1:30410 [upload_job.py:push():130] Uploaded file /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_190333-82mnef5m/files/output.log +2024-06-08 20:15:56,213 INFO wandb-upload_3:30410 [upload_job.py:push():130] Uploaded file /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_190333-82mnef5m/files/wandb-summary.json +2024-06-08 20:15:56,242 INFO wandb-upload_2:30410 [upload_job.py:push():130] Uploaded file /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_190333-82mnef5m/files/requirements.txt +2024-06-08 20:15:56,443 INFO Thread-11 (_thread_body):30410 [sender.py:transition_state():614] send defer: 11 +2024-06-08 20:15:56,443 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: defer +2024-06-08 20:15:56,443 INFO HandlerThread:30410 [handler.py:handle_request_defer():184] handle defer: 11 +2024-06-08 20:15:56,443 DEBUG SenderThread:30410 [sender.py:send_request():406] send_request: defer +2024-06-08 20:15:56,443 INFO SenderThread:30410 [sender.py:send_request_defer():610] handle sender defer: 11 +2024-06-08 20:15:56,443 INFO SenderThread:30410 [file_pusher.py:join():175] waiting for file pusher +2024-06-08 20:15:56,444 INFO SenderThread:30410 [sender.py:transition_state():614] send defer: 12 +2024-06-08 20:15:56,444 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: defer +2024-06-08 20:15:56,444 INFO HandlerThread:30410 [handler.py:handle_request_defer():184] handle defer: 12 +2024-06-08 20:15:56,444 DEBUG SenderThread:30410 [sender.py:send_request():406] send_request: defer +2024-06-08 20:15:56,444 INFO SenderThread:30410 [sender.py:send_request_defer():610] handle sender defer: 12 +2024-06-08 20:15:56,444 INFO SenderThread:30410 [file_stream.py:finish():601] file stream finish called +2024-06-08 20:15:56,734 INFO SenderThread:30410 [file_stream.py:finish():605] file stream finish is done +2024-06-08 20:15:56,734 INFO SenderThread:30410 [sender.py:transition_state():614] send defer: 13 +2024-06-08 20:15:56,734 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: defer +2024-06-08 20:15:56,734 INFO HandlerThread:30410 [handler.py:handle_request_defer():184] handle defer: 13 +2024-06-08 20:15:56,735 DEBUG SenderThread:30410 [sender.py:send_request():406] send_request: defer +2024-06-08 20:15:56,735 INFO SenderThread:30410 [sender.py:send_request_defer():610] handle sender defer: 13 +2024-06-08 20:15:56,735 INFO SenderThread:30410 [sender.py:transition_state():614] send defer: 14 +2024-06-08 20:15:56,735 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: defer +2024-06-08 20:15:56,735 INFO HandlerThread:30410 [handler.py:handle_request_defer():184] handle defer: 14 +2024-06-08 20:15:56,735 DEBUG SenderThread:30410 [sender.py:send():379] send: final +2024-06-08 20:15:56,735 DEBUG SenderThread:30410 [sender.py:send():379] send: footer +2024-06-08 20:15:56,735 DEBUG SenderThread:30410 [sender.py:send_request():406] send_request: defer +2024-06-08 20:15:56,735 INFO SenderThread:30410 [sender.py:send_request_defer():610] handle sender defer: 14 +2024-06-08 20:15:56,736 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: poll_exit +2024-06-08 20:15:56,736 DEBUG SenderThread:30410 [sender.py:send_request():406] send_request: poll_exit +2024-06-08 20:15:56,736 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 20:15:56,737 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: get_summary +2024-06-08 20:15:56,738 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: sampled_history +2024-06-08 20:15:56,739 DEBUG HandlerThread:30410 [handler.py:handle_request():158] handle_request: shutdown +2024-06-08 20:15:56,739 INFO HandlerThread:30410 [handler.py:finish():882] shutting down handler +2024-06-08 20:15:57,736 INFO WriterThread:30410 [datastore.py:close():296] close: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_190333-82mnef5m/run-82mnef5m.wandb +2024-06-08 20:15:57,736 INFO SenderThread:30410 [sender.py:finish():1608] shutting down sender +2024-06-08 20:15:57,736 INFO SenderThread:30410 [file_pusher.py:finish():169] shutting down file pusher +2024-06-08 20:15:57,736 INFO SenderThread:30410 [file_pusher.py:join():175] waiting for file pusher diff --git a/lm-evaluation-harness/wandb/debug.log b/lm-evaluation-harness/wandb/debug.log new file mode 100644 index 0000000000000000000000000000000000000000..4118681593446b0aac887d9eec84ee3c040350ba --- /dev/null +++ b/lm-evaluation-harness/wandb/debug.log @@ -0,0 +1,36 @@ +2024-06-08 19:03:33,772 INFO MainThread:30255 [wandb_setup.py:_flush():76] Current SDK version is 0.17.1 +2024-06-08 19:03:33,772 INFO MainThread:30255 [wandb_setup.py:_flush():76] Configure stats pid to 30255 +2024-06-08 19:03:33,772 INFO MainThread:30255 [wandb_setup.py:_flush():76] Loading settings from /root/.config/wandb/settings +2024-06-08 19:03:33,772 INFO MainThread:30255 [wandb_setup.py:_flush():76] Loading settings from /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/settings +2024-06-08 19:03:33,773 INFO MainThread:30255 [wandb_setup.py:_flush():76] Loading settings from environment variables: {} +2024-06-08 19:03:33,773 INFO MainThread:30255 [wandb_setup.py:_flush():76] Applying setup settings: {'_disable_service': False} +2024-06-08 19:03:33,773 WARNING MainThread:30255 [wandb_setup.py:_flush():76] Could not find program at -m lm_eval.__main__ +2024-06-08 19:03:33,773 INFO MainThread:30255 [wandb_setup.py:_flush():76] Inferring run settings from compute environment: {'program_relpath': None, 'program': '-m lm_eval.__main__'} +2024-06-08 19:03:33,773 INFO MainThread:30255 [wandb_setup.py:_flush():76] Applying login settings: {} +2024-06-08 19:03:33,773 INFO MainThread:30255 [wandb_init.py:_log_setup():520] Logging user logs to /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_190333-82mnef5m/logs/debug.log +2024-06-08 19:03:33,773 INFO MainThread:30255 [wandb_init.py:_log_setup():521] Logging internal logs to /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_190333-82mnef5m/logs/debug-internal.log +2024-06-08 19:03:33,773 INFO MainThread:30255 [wandb_init.py:init():560] calling init triggers +2024-06-08 19:03:33,773 INFO MainThread:30255 [wandb_init.py:init():567] wandb.init called with sweep_config: {} +config: {} +2024-06-08 19:03:33,773 INFO MainThread:30255 [wandb_init.py:init():610] starting backend +2024-06-08 19:03:33,773 INFO MainThread:30255 [wandb_init.py:init():614] setting up manager +2024-06-08 19:03:33,776 INFO MainThread:30255 [backend.py:_multiprocessing_setup():105] multiprocessing start_methods=fork,spawn,forkserver, using: spawn +2024-06-08 19:03:33,777 INFO MainThread:30255 [wandb_init.py:init():622] backend started and connected +2024-06-08 19:03:33,781 INFO MainThread:30255 [wandb_init.py:init():711] updated telemetry +2024-06-08 19:03:33,790 INFO MainThread:30255 [wandb_init.py:init():744] communicating run to backend with 90.0 second timeout +2024-06-08 19:03:33,992 INFO MainThread:30255 [wandb_run.py:_on_init():2402] communicating current version +2024-06-08 19:03:34,052 INFO MainThread:30255 [wandb_run.py:_on_init():2411] got version response +2024-06-08 19:03:34,052 INFO MainThread:30255 [wandb_init.py:init():795] starting run threads in backend +2024-06-08 19:03:34,350 INFO MainThread:30255 [wandb_run.py:_console_start():2380] atexit reg +2024-06-08 19:03:34,350 INFO MainThread:30255 [wandb_run.py:_redirect():2235] redirect: wrap_raw +2024-06-08 19:03:34,351 INFO MainThread:30255 [wandb_run.py:_redirect():2300] Wrapping output streams. +2024-06-08 19:03:34,351 INFO MainThread:30255 [wandb_run.py:_redirect():2325] Redirects installed. +2024-06-08 19:03:34,353 INFO MainThread:30255 [wandb_init.py:init():838] run started, returning control to user process +2024-06-08 20:15:51,241 INFO MainThread:30255 [wandb_run.py:_config_callback():1382] config_cb None None {'task_configs': {'arc_easy': {'task': 'arc_easy', 'group': ['ai2_arc'], 'dataset_path': 'allenai/ai2_arc', 'dataset_name': 'ARC-Easy', 'training_split': 'train', 'validation_split': 'validation', 'test_split': 'test', 'doc_to_text': 'Question: {{question}}\nAnswer:', 'doc_to_target': '{{choices.label.index(answerKey)}}', 'doc_to_choice': '{{choices.text}}', 'description': '', 'target_delimiter': ' ', 'fewshot_delimiter': '\n\n', 'num_fewshot': 0, 'metric_list': [{'metric': 'acc', 'aggregation': 'mean', 'higher_is_better': True}, {'metric': 'acc_norm', 'aggregation': 'mean', 'higher_is_better': True}], 'output_type': 'multiple_choice', 'repeats': 1, 'should_decontaminate': True, 'doc_to_decontamination_query': 'Question: {{question}}\nAnswer:', 'metadata': {'version': 1.0}}, 'boolq': {'task': 'boolq', 'group': ['super-glue-lm-eval-v1'], 'dataset_path': 'super_glue', 'dataset_name': 'boolq', 'training_split': 'train', 'validation_split': 'validation', 'doc_to_text': '{{passage}}\nQuestion: {{question}}?\nAnswer:', 'doc_to_target': 'label', 'doc_to_choice': ['no', 'yes'], 'description': '', 'target_delimiter': ' ', 'fewshot_delimiter': '\n\n', 'num_fewshot': 0, 'metric_list': [{'metric': 'acc'}], 'output_type': 'multiple_choice', 'repeats': 1, 'should_decontaminate': True, 'doc_to_decontamination_query': 'passage', 'metadata': {'version': 2.0}}, 'copa': {'task': 'copa', 'group': ['super-glue-lm-eval-v1'], 'dataset_path': 'super_glue', 'dataset_name': 'copa', 'training_split': 'train', 'validation_split': 'validation', 'doc_to_text': 'def doc_to_text(doc):\n # Drop the period\n connector = {\n "cause": "because",\n "effect": "therefore",\n }[doc["question"]]\n return doc["premise"].strip()[:-1] + f" {connector}"\n', 'doc_to_target': 'def doc_to_target(doc):\n correct_choice = doc["choice1"] if doc["label"] == 0 else doc["choice2"]\n # Connect the sentences\n return " " + convert_choice(correct_choice)\n', 'doc_to_choice': 'def doc_to_choice(doc):\n return [" " + convert_choice(doc["choice1"]), " " + convert_choice(doc["choice2"])]\n', 'description': '', 'target_delimiter': ' ', 'fewshot_delimiter': '\n\n', 'num_fewshot': 0, 'metric_list': [{'metric': 'acc'}], 'output_type': 'multiple_choice', 'repeats': 1, 'should_decontaminate': False, 'metadata': {'version': 1.0}}, 'indic_arc_challenge_hi': {'task': 'indic_arc_challenge_hi', 'group': 'Cognitive-Lab/Indic-ARC-Challenge', 'dataset_path': 'Cognitive-Lab/Indic-ARC-Challenge', 'dataset_name': 'hi', 'test_split': 'test', 'doc_to_text': 'Question: {{translated_question}}\nAnswer:', 'doc_to_target': '{{translated_choices.label.index(answerKey)}}', 'doc_to_choice': '{{translated_choices.text}}', 'description': '', 'target_delimiter': ' ', 'fewshot_delimiter': '\n\n', 'num_fewshot': 0, 'metric_list': [{'metric': 'acc', 'aggregation': 'mean', 'higher_is_better': True}], 'output_type': 'multiple_choice', 'repeats': 1, 'should_decontaminate': True, 'doc_to_decontamination_query': 'Question: {{translated_question}}\nAnswer:', 'metadata': {'version': 1.0}}, 'indic_arc_easy_hi': {'task': 'indic_arc_easy_hi', 'group': 'Cognitive-Lab/Indic-ARC-Easy', 'dataset_path': 'Cognitive-Lab/Indic-ARC-Easy', 'dataset_name': 'hi', 'test_split': 'test', 'doc_to_text': 'Question: {{translated_question}}\nAnswer:', 'doc_to_target': '{{translated_choices.label.index(answerKey)}}', 'doc_to_choice': '{{translated_choices.text}}', 'description': '', 'target_delimiter': ' ', 'fewshot_delimiter': '\n\n', 'num_fewshot': 0, 'metric_list': [{'metric': 'acc', 'aggregation': 'mean', 'higher_is_better': True}], 'output_type': 'multiple_choice', 'repeats': 1, 'should_decontaminate': True, 'doc_to_decontamination_query': 'Question: {{translated_question}}\nAnswer:', 'metadata': {'version': 1.0}}, 'indic_boolq_hi': {'task': 'indic_boolq_hi', 'group': 'Cognitive-Lab/Indic-BoolQ', 'dataset_path': 'Cognitive-Lab/Indic-BoolQ', 'dataset_name': 'hi', 'validation_split': 'validation', 'doc_to_text': 'Passage: {translated_passage}\nQuestion: {translated_question.strip()}\nAnswer:', 'doc_to_target': 'answer', 'doc_to_choice': ['true', 'false'], 'description': '', 'target_delimiter': ' ', 'fewshot_delimiter': '\n\n', 'num_fewshot': 0, 'metric_list': [{'metric': 'acc', 'aggregation': 'mean', 'higher_is_better': True}], 'output_type': 'multiple_choice', 'repeats': 1, 'should_decontaminate': False, 'metadata': {'version': 1.0}}, 'mrpc': {'task': 'mrpc', 'group': 'glue', 'dataset_path': 'glue', 'dataset_name': 'mrpc', 'training_split': 'train', 'validation_split': 'validation', 'doc_to_text': 'Sentence 1: {{sentence1}}\nSentence 2: {{sentence2}}\nQuestion: Do both sentences mean the same thing?\nAnswer:', 'doc_to_target': 'label', 'doc_to_choice': ['no', 'yes'], 'description': '', 'target_delimiter': ' ', 'fewshot_delimiter': '\n\n', 'num_fewshot': 0, 'metric_list': [{'metric': 'acc'}, {'metric': 'f1'}], 'output_type': 'multiple_choice', 'repeats': 1, 'should_decontaminate': False, 'metadata': {'version': 1.0}}, 'piqa': {'task': 'piqa', 'dataset_path': 'piqa', 'training_split': 'train', 'validation_split': 'validation', 'doc_to_text': 'Question: {{goal}}\nAnswer:', 'doc_to_target': 'label', 'doc_to_choice': '{{[sol1, sol2]}}', 'description': '', 'target_delimiter': ' ', 'fewshot_delimiter': '\n\n', 'num_fewshot': 0, 'metric_list': [{'metric': 'acc', 'aggregation': 'mean', 'higher_is_better': True}, {'metric': 'acc_norm', 'aggregation': 'mean', 'higher_is_better': True}], 'output_type': 'multiple_choice', 'repeats': 1, 'should_decontaminate': True, 'doc_to_decontamination_query': 'goal', 'metadata': {'version': 1.0}}, 'sst2': {'task': 'sst2', 'group': 'glue', 'dataset_path': 'glue', 'dataset_name': 'sst2', 'training_split': 'train', 'validation_split': 'validation', 'doc_to_text': '{{sentence}}\nQuestion: Is this sentence positive or negative?\nAnswer:', 'doc_to_target': 'label', 'doc_to_choice': ['negative', 'positive'], 'description': '', 'target_delimiter': ' ', 'fewshot_delimiter': '\n\n', 'num_fewshot': 0, 'metric_list': [{'metric': 'acc'}], 'output_type': 'multiple_choice', 'repeats': 1, 'should_decontaminate': False, 'metadata': {'version': 1.0}}, 'winogrande': {'task': 'winogrande', 'dataset_path': 'winogrande', 'dataset_name': 'winogrande_xl', 'training_split': 'train', 'validation_split': 'validation', 'doc_to_text': 'def doc_to_text(doc):\n answer_to_num = {"1": 0, "2": 1}\n return answer_to_num[doc["answer"]]\n', 'doc_to_target': 'def doc_to_target(doc):\n idx = doc["sentence"].index("_") + 1\n return doc["sentence"][idx:].strip()\n', 'doc_to_choice': 'def doc_to_choice(doc):\n idx = doc["sentence"].index("_")\n options = [doc["option1"], doc["option2"]]\n return [doc["sentence"][:idx] + opt for opt in options]\n', 'description': '', 'target_delimiter': ' ', 'fewshot_delimiter': '\n\n', 'num_fewshot': 0, 'metric_list': [{'metric': 'acc', 'aggregation': 'mean', 'higher_is_better': True}], 'output_type': 'multiple_choice', 'repeats': 1, 'should_decontaminate': True, 'doc_to_decontamination_query': 'sentence', 'metadata': {'version': 1.0}}}, 'cli_configs': {'model': 'hf', 'model_args': 'pretrained=/mnt/weka/peacock/experiments/llama/eval/checkpoint-enhibn-updated/llamav2-3b/hf/global_step240000,tokenizer=/mnt/weka/peacock/tokenization/trained-tokenizer/enhiben_50k_hf/ConvertedTokenizer', 'batch_size': 'auto', 'batch_sizes': [64], 'device': None, 'use_cache': None, 'limit': None, 'bootstrap_iters': 100000, 'gen_kwargs': None}} +2024-06-08 20:15:51,834 INFO MainThread:30255 [wandb_run.py:_finish():2109] finishing run smlgenai/engl-hi-eval/82mnef5m +2024-06-08 20:15:51,834 INFO MainThread:30255 [wandb_run.py:_atexit_cleanup():2349] got exitcode: 0 +2024-06-08 20:15:51,835 INFO MainThread:30255 [wandb_run.py:_restore():2332] restore +2024-06-08 20:15:51,835 INFO MainThread:30255 [wandb_run.py:_restore():2338] restore done +2024-06-08 20:15:57,740 INFO MainThread:30255 [wandb_run.py:_footer_history_summary_info():4008] rendering history +2024-06-08 20:15:57,741 INFO MainThread:30255 [wandb_run.py:_footer_history_summary_info():4040] rendering summary +2024-06-08 20:15:57,748 INFO MainThread:30255 [wandb_run.py:_footer_sync_info():3967] logging synced files diff --git a/lm-evaluation-harness/wandb/run-20240523_130326-1oo0voi6/files/config.yaml b/lm-evaluation-harness/wandb/run-20240523_130326-1oo0voi6/files/config.yaml new file mode 100644 index 0000000000000000000000000000000000000000..bcc5d53e566ea9d86da4d02a79606c972f82b99c --- /dev/null +++ b/lm-evaluation-harness/wandb/run-20240523_130326-1oo0voi6/files/config.yaml @@ -0,0 +1,43 @@ +wandb_version: 1 + +_wandb: + desc: null + value: + python_version: 3.10.12 + cli_version: 0.17.0 + framework: huggingface + huggingface_version: 4.41.1 + is_jupyter_run: false + is_kaggle_kernel: false + start_time: 1716469406 + t: + 1: + - 1 + - 5 + - 11 + - 49 + - 51 + - 53 + - 55 + - 71 + - 98 + - 100 + 2: + - 1 + - 5 + - 11 + - 49 + - 51 + - 53 + - 55 + - 71 + - 98 + - 100 + 3: + - 23 + 4: 3.10.12 + 5: 0.17.0 + 6: 4.41.1 + 8: + - 5 + 13: linux-x86_64 diff --git a/lm-evaluation-harness/wandb/run-20240523_130326-1oo0voi6/files/output.log b/lm-evaluation-harness/wandb/run-20240523_130326-1oo0voi6/files/output.log new file mode 100644 index 0000000000000000000000000000000000000000..4cd709d32b0fd0a72625c84620aed9da0e7ed395 --- /dev/null +++ b/lm-evaluation-harness/wandb/run-20240523_130326-1oo0voi6/files/output.log @@ -0,0 +1,34 @@ + +2024-05-23:13:03:27,131 INFO [__main__.py:251] Verbosity set to INFO +2024-05-23:13:03:35,583 INFO [__main__.py:335] Selected Tasks: ['arc_easy', 'hellaswag', 'mrpc', 'openbookqa', 'sst2', 'winogrande'] +2024-05-23:13:03:35,584 INFO [evaluator.py:131] Setting random seed to 0 | Setting numpy seed to 1234 | Setting torch manual seed to 1234 +2024-05-23:13:03:35,584 INFO [evaluator.py:177] Initializing hf model, with arguments: {'pretrained': '/mnt/weka/peacock/experiments/llama/checkpoint/llamav2-3b//hf_ckpt//global_step10000'} +2024-05-23:13:03:37,878 INFO [huggingface.py:164] Using device 'cuda' +Traceback (most recent call last): + File "/usr/lib/python3.10/runpy.py", line 196, in _run_module_as_main + return _run_code(code, main_globals, None, + File "/usr/lib/python3.10/runpy.py", line 86, in _run_code + exec(code, run_globals) + File "/mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/lm_eval/__main__.py", line 417, in + cli_evaluate() + File "/mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/lm_eval/__main__.py", line 341, in cli_evaluate + results = evaluator.simple_evaluate( + File "/mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/lm_eval/utils.py", line 288, in _wrapper + return fn(*args, **kwargs) + File "/mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/lm_eval/evaluator.py", line 180, in simple_evaluate + lm = lm_eval.api.registry.get_model(model).create_from_arg_string( + File "/mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/lm_eval/api/model.py", line 134, in create_from_arg_string + return cls(**args, **args2) + File "/mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/lm_eval/models/huggingface.py", line 190, in __init__ + self._get_config( + File "/mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/lm_eval/models/huggingface.py", line 471, in _get_config + self._config = transformers.AutoConfig.from_pretrained( + File "/usr/local/lib/python3.10/dist-packages/transformers/models/auto/configuration_auto.py", line 934, in from_pretrained + config_dict, unused_kwargs = PretrainedConfig.get_config_dict(pretrained_model_name_or_path, **kwargs) + File "/usr/local/lib/python3.10/dist-packages/transformers/configuration_utils.py", line 632, in get_config_dict + config_dict, kwargs = cls._get_config_dict(pretrained_model_name_or_path, **kwargs) + File "/usr/local/lib/python3.10/dist-packages/transformers/configuration_utils.py", line 689, in _get_config_dict + resolved_config_file = cached_file( + File "/usr/local/lib/python3.10/dist-packages/transformers/utils/hub.py", line 370, in cached_file + raise EnvironmentError( +OSError: /mnt/weka/peacock/experiments/llama/checkpoint/llamav2-3b//hf_ckpt//global_step10000 does not appear to have a file named config.json. Checkout 'https://huggingface.co//mnt/weka/peacock/experiments/llama/checkpoint/llamav2-3b//hf_ckpt//global_step10000/tree/main' for available files. \ No newline at end of file diff --git a/lm-evaluation-harness/wandb/run-20240523_130326-1oo0voi6/files/requirements.txt b/lm-evaluation-harness/wandb/run-20240523_130326-1oo0voi6/files/requirements.txt new file mode 100644 index 0000000000000000000000000000000000000000..f675c3016b5332c1acf28f436e0b60adeead9c12 --- /dev/null +++ b/lm-evaluation-harness/wandb/run-20240523_130326-1oo0voi6/files/requirements.txt @@ -0,0 +1,155 @@ +DataProperty==1.0.1 +GitPython==3.1.43 +Jinja2==3.1.4 +Markdown==3.6 +MarkupSafe==2.1.5 +Pillow-SIMD==7.0.0.post3 +PyYAML==6.0 +Werkzeug==3.0.3 +absl-py==2.1.0 +accelerate==0.30.1 +aiohttp==3.9.5 +aiosignal==1.3.1 +async-timeout==4.0.3 +attrs==23.2.0 +av==9.2.0 +cachetools==5.3.3 +certifi==2024.2.2 +cffi==1.15.1 +cfgv==3.4.0 +chardet==5.2.0 +charset-normalizer==3.3.2 +click==8.1.7 +cmake==3.29.2 +colorama==0.4.6 +datasets==2.19.1 +deepspeed==0.12.4+hpu.synapse.v1.15.1 +dill==0.3.8 +distlib==0.3.8 +docker-pycreds==0.4.0 +einops==0.8.0 +evaluate==0.4.2 +exceptiongroup==1.2.1 +expecttest==0.2.1 +filelock==3.14.0 +frozenlist==1.4.1 +fsspec==2024.3.1 +gitdb==4.0.11 +google-auth-oauthlib==0.4.6 +google-auth==2.29.0 +grpcio==1.63.0 +habana-media-loader==1.15.1.15 +habana-pyhlml==1.15.1.15 +habana-torch-dataloader==1.15.1.15 +habana-torch-plugin==1.15.1.15 +habana_gpu_migration==1.15.1.15 +habana_quantization_toolkit==1.15.1.15 +hjson==3.1.0 +huggingface-hub==0.23.1 +identify==2.5.36 +idna==3.7 +iniconfig==2.0.0 +joblib==1.4.2 +jsonlines==4.0.0 +lightning-habana==1.4.0 +lightning-utilities==0.11.2 +lightning==2.2.0.post0 +lm_eval==0.4.2 +lm_eval==0.4.2 +lm_eval==0.4.2 +lxml==5.2.2 +mbstrdecoder==1.1.3 +more-itertools==10.2.0 +mpi4py==3.1.4 +mpmath==1.3.0 +multidict==6.0.5 +multiprocess==0.70.16 +networkx==3.3 +ninja==1.11.1.1 +nltk==3.8.1 +nodeenv==1.8.0 +numexpr==2.10.0 +numpy==1.23.5 +oauthlib==3.2.2 +packaging==24.0 +pandas==2.0.1 +pathspec==0.12.1 +pathvalidate==3.2.0 +peft==0.11.1 +perfetto==0.7.0 +pillow==10.3.0 +pip==22.0.2 +pip==23.3.1 +platformdirs==4.2.1 +pluggy==1.5.0 +portalocker==2.8.2 +pre-commit==3.3.3 +pretty-errors==1.2.25 +protobuf==3.20.3 +psutil==5.9.8 +py-cpuinfo==9.0.0 +pyarrow-hotfix==0.6 +pyarrow==16.1.0 +pyasn1==0.6.0 +pyasn1_modules==0.4.0 +pybind11==2.10.4 +pycparser==2.22 +pydantic==1.10.13 +pynvml==8.0.4 +pytablewriter==1.2.0 +pytest==8.2.0 +python-dateutil==2.9.0.post0 +pytorch-lightning==2.2.4 +pytz==2024.1 +regex==2023.5.5 +requests-oauthlib==2.0.0 +requests==2.31.0 +rouge_score==0.1.2 +rsa==4.9 +sacrebleu==2.4.2 +safetensors==0.4.3 +scikit-learn==1.5.0 +scipy==1.13.1 +sentencepiece==0.2.0 +sentry-sdk==2.3.0 +setproctitle==1.3.3 +setuptools==59.6.0 +setuptools==69.5.1 +six==1.16.0 +smmap==5.0.1 +sqlitedict==2.1.0 +symengine==0.11.0 +sympy==1.12 +tabledata==1.3.3 +tabulate==0.9.0 +tcolorpy==0.1.6 +tdqm==0.0.1 +tensorboard-data-server==0.6.1 +tensorboard-plugin-wit==1.8.1 +tensorboard==2.11.2 +threadpoolctl==3.5.0 +tokenizers==0.19.1 +tomli==2.0.1 +torch==2.2.0a0+git8964477 +torch_tb_profiler==0.4.0 +torchaudio==2.2.0+08901ad +torchdata==0.7.1+5e6f7b7 +torchmetrics==1.4.0 +torchtext==0.17.0+400da5c +torchvision==0.17.0+b2383d4 +tqdm-multiprocess==0.0.11 +tqdm==4.66.4 +transformers==4.41.1 +typepy==1.3.2 +typing_extensions==4.11.0 +tzdata==2024.1 +urllib3==1.26.18 +virtualenv==20.26.1 +wandb==0.17.0 +wheel==0.37.1 +wheel==0.43.0 +word2number==1.1 +xxhash==3.4.1 +yamllint==1.35.1 +yarl==1.9.4 +zstandard==0.22.0 \ No newline at end of file diff --git a/lm-evaluation-harness/wandb/run-20240523_130326-1oo0voi6/files/wandb-metadata.json b/lm-evaluation-harness/wandb/run-20240523_130326-1oo0voi6/files/wandb-metadata.json new file mode 100644 index 0000000000000000000000000000000000000000..b1ea55ee37555d0273e75cf9834f43a294b343c2 --- /dev/null +++ b/lm-evaluation-harness/wandb/run-20240523_130326-1oo0voi6/files/wandb-metadata.json @@ -0,0 +1,850 @@ +{ + "os": "Linux-5.15.0-92-generic-x86_64-with-glibc2.35", + "python": "3.10.12", + "heartbeatAt": "2024-05-23T13:03:26.924960", + "startedAt": "2024-05-23T13:03:26.400431", + "docker": null, + "cuda": null, + "args": [ + "--model", + "hf", + "--model_args", + "pretrained=/mnt/weka/peacock/experiments/llama/checkpoint/llamav2-3b//hf_ckpt//global_step10000", + "--tasks", + "hellaswag,arc_easy,openbookqa,winogrande,sst2,mrpc", + "--batch_size", + "auto", + "--wandb_args", + "project=bharatgpt,group=trial_expt_2" + ], + "state": "running", + "program": "-m lm_eval.__main__", + "codePathLocal": null, + "git": { + "remote": "https://github.com/EleutherAI/lm-evaluation-harness", + "commit": null + }, + "email": null, + "root": "/mnt/weka/peacock/idc/cronscript/lm-evaluation-harness", + "host": "peacock-evaluation-worker-0", + "username": "root", + "executable": "/usr/bin/python3", + "cpu_count": 80, + "cpu_count_logical": 160, + "cpu_freq": { + "current": 2332.72725, + "min": 800.0, + "max": 3400.0 + }, + "cpu_freq_per_core": [ + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3198.822, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + } + ], + "disk": { + "/": { + "total": 877.6341285705566, + "used": 211.61690521240234 + } + }, + "memory": { + "total": 1007.4379539489746 + } +} diff --git a/lm-evaluation-harness/wandb/run-20240523_130326-1oo0voi6/files/wandb-summary.json b/lm-evaluation-harness/wandb/run-20240523_130326-1oo0voi6/files/wandb-summary.json new file mode 100644 index 0000000000000000000000000000000000000000..8bf99d152ad35c3699ec8600ecb8b169d4e35875 --- /dev/null +++ b/lm-evaluation-harness/wandb/run-20240523_130326-1oo0voi6/files/wandb-summary.json @@ -0,0 +1 @@ +{"_wandb": {"runtime": 11}} \ No newline at end of file diff --git a/lm-evaluation-harness/wandb/run-20240523_130326-1oo0voi6/logs/debug-internal.log b/lm-evaluation-harness/wandb/run-20240523_130326-1oo0voi6/logs/debug-internal.log new file mode 100644 index 0000000000000000000000000000000000000000..58073a41c284165be87895395d88177b8ef3fb30 --- /dev/null +++ b/lm-evaluation-harness/wandb/run-20240523_130326-1oo0voi6/logs/debug-internal.log @@ -0,0 +1,183 @@ +2024-05-23 13:03:26,421 INFO StreamThr :1072 [internal.py:wandb_internal():85] W&B internal server running at pid: 1072, started at: 2024-05-23 13:03:26.419743 +2024-05-23 13:03:26,426 DEBUG HandlerThread:1072 [handler.py:handle_request():158] handle_request: status +2024-05-23 13:03:26,427 INFO WriterThread:1072 [datastore.py:open_for_write():87] open: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240523_130326-1oo0voi6/run-1oo0voi6.wandb +2024-05-23 13:03:26,429 DEBUG SenderThread:1072 [sender.py:send():378] send: header +2024-05-23 13:03:26,433 DEBUG SenderThread:1072 [sender.py:send():378] send: run +2024-05-23 13:03:26,724 INFO SenderThread:1072 [dir_watcher.py:__init__():211] watching files in: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240523_130326-1oo0voi6/files +2024-05-23 13:03:26,725 INFO SenderThread:1072 [sender.py:_start_run_threads():1123] run started: 1oo0voi6 with start time 1716469406.420894 +2024-05-23 13:03:26,726 DEBUG HandlerThread:1072 [handler.py:handle_request():158] handle_request: check_version +2024-05-23 13:03:26,726 DEBUG SenderThread:1072 [sender.py:send_request():405] send_request: check_version +2024-05-23 13:03:26,850 DEBUG HandlerThread:1072 [handler.py:handle_request():158] handle_request: run_start +2024-05-23 13:03:26,852 DEBUG HandlerThread:1072 [system_info.py:__init__():26] System info init +2024-05-23 13:03:26,852 DEBUG HandlerThread:1072 [system_info.py:__init__():41] System info init done +2024-05-23 13:03:26,852 INFO HandlerThread:1072 [system_monitor.py:start():194] Starting system monitor +2024-05-23 13:03:26,852 INFO SystemMonitor:1072 [system_monitor.py:_start():158] Starting system asset monitoring threads +2024-05-23 13:03:26,853 INFO HandlerThread:1072 [system_monitor.py:probe():214] Collecting system info +2024-05-23 13:03:26,859 INFO SystemMonitor:1072 [interfaces.py:start():188] Started cpu monitoring +2024-05-23 13:03:26,860 INFO SystemMonitor:1072 [interfaces.py:start():188] Started disk monitoring +2024-05-23 13:03:26,860 INFO SystemMonitor:1072 [interfaces.py:start():188] Started memory monitoring +2024-05-23 13:03:26,860 INFO SystemMonitor:1072 [interfaces.py:start():188] Started network monitoring +2024-05-23 13:03:26,924 DEBUG HandlerThread:1072 [system_info.py:probe():150] Probing system +2024-05-23 13:03:26,928 DEBUG HandlerThread:1072 [system_info.py:_probe_git():135] Probing git +2024-05-23 13:03:26,938 ERROR HandlerThread:1072 [gitlib.py:root():92] git root error: Cmd('git') failed due to: exit code(128) + cmdline: git rev-parse --show-toplevel + stderr: 'fatal: detected dubious ownership in repository at '/mnt/weka/peacock/idc/cronscript/lm-evaluation-harness' +To add an exception for this directory, call: + + git config --global --add safe.directory /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness' +2024-05-23 13:03:26,938 DEBUG HandlerThread:1072 [system_info.py:_probe_git():143] Probing git done +2024-05-23 13:03:26,938 DEBUG HandlerThread:1072 [system_info.py:probe():198] Probing system done +2024-05-23 13:03:26,938 DEBUG HandlerThread:1072 [system_monitor.py:probe():223] {'os': 'Linux-5.15.0-92-generic-x86_64-with-glibc2.35', 'python': '3.10.12', 'heartbeatAt': '2024-05-23T13:03:26.924960', 'startedAt': '2024-05-23T13:03:26.400431', 'docker': None, 'cuda': None, 'args': ('--model', 'hf', '--model_args', 'pretrained=/mnt/weka/peacock/experiments/llama/checkpoint/llamav2-3b//hf_ckpt//global_step10000', '--tasks', 'hellaswag,arc_easy,openbookqa,winogrande,sst2,mrpc', '--batch_size', 'auto', '--wandb_args', 'project=bharatgpt,group=trial_expt_2'), 'state': 'running', 'program': '-m lm_eval.__main__', 'codePathLocal': None, 'git': {'remote': 'https://github.com/EleutherAI/lm-evaluation-harness', 'commit': None}, 'email': None, 'root': '/mnt/weka/peacock/idc/cronscript/lm-evaluation-harness', 'host': 'peacock-evaluation-worker-0', 'username': 'root', 'executable': '/usr/bin/python3', 'cpu_count': 80, 'cpu_count_logical': 160, 'cpu_freq': {'current': 2332.72725, 'min': 800.0, 'max': 3400.0}, 'cpu_freq_per_core': [{'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 3198.822, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}], 'disk': {'/': {'total': 877.6341285705566, 'used': 211.61690521240234}}, 'memory': {'total': 1007.4379539489746}} +2024-05-23 13:03:26,938 INFO HandlerThread:1072 [system_monitor.py:probe():224] Finished collecting system info +2024-05-23 13:03:26,938 INFO HandlerThread:1072 [system_monitor.py:probe():227] Publishing system info +2024-05-23 13:03:26,941 INFO HandlerThread:1072 [system_monitor.py:probe():229] Finished publishing system info +2024-05-23 13:03:26,946 DEBUG SenderThread:1072 [sender.py:send():378] send: files +2024-05-23 13:03:26,946 INFO SenderThread:1072 [sender.py:_save_file():1389] saving file wandb-metadata.json with policy now +2024-05-23 13:03:27,125 DEBUG HandlerThread:1072 [handler.py:handle_request():158] handle_request: python_packages +2024-05-23 13:03:27,125 DEBUG SenderThread:1072 [sender.py:send_request():405] send_request: python_packages +2024-05-23 13:03:27,126 DEBUG HandlerThread:1072 [handler.py:handle_request():158] handle_request: stop_status +2024-05-23 13:03:27,127 DEBUG SenderThread:1072 [sender.py:send_request():405] send_request: stop_status +2024-05-23 13:03:27,218 DEBUG SenderThread:1072 [sender.py:send():378] send: telemetry +2024-05-23 13:03:27,522 INFO wandb-upload_0:1072 [upload_job.py:push():130] Uploaded file /tmp/tmpzbrvci8cwandb/r70fz28y-wandb-metadata.json +2024-05-23 13:03:27,725 INFO Thread-12 :1072 [dir_watcher.py:_on_file_created():271] file/dir created: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240523_130326-1oo0voi6/files/wandb-metadata.json +2024-05-23 13:03:27,725 INFO Thread-12 :1072 [dir_watcher.py:_on_file_created():271] file/dir created: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240523_130326-1oo0voi6/files/output.log +2024-05-23 13:03:27,726 INFO Thread-12 :1072 [dir_watcher.py:_on_file_created():271] file/dir created: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240523_130326-1oo0voi6/files/requirements.txt +2024-05-23 13:03:29,725 INFO Thread-12 :1072 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240523_130326-1oo0voi6/files/output.log +2024-05-23 13:03:32,223 DEBUG HandlerThread:1072 [handler.py:handle_request():158] handle_request: status_report +2024-05-23 13:03:37,585 DEBUG HandlerThread:1072 [handler.py:handle_request():158] handle_request: status_report +2024-05-23 13:03:37,732 INFO Thread-12 :1072 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240523_130326-1oo0voi6/files/output.log +2024-05-23 13:03:37,887 DEBUG SenderThread:1072 [sender.py:send():378] send: exit +2024-05-23 13:03:37,887 INFO SenderThread:1072 [sender.py:send_exit():585] handling exit code: 1 +2024-05-23 13:03:37,887 INFO SenderThread:1072 [sender.py:send_exit():587] handling runtime: 11 +2024-05-23 13:03:37,889 INFO SenderThread:1072 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end +2024-05-23 13:03:37,889 INFO SenderThread:1072 [sender.py:send_exit():593] send defer +2024-05-23 13:03:37,889 DEBUG HandlerThread:1072 [handler.py:handle_request():158] handle_request: defer +2024-05-23 13:03:37,889 INFO HandlerThread:1072 [handler.py:handle_request_defer():184] handle defer: 0 +2024-05-23 13:03:37,890 DEBUG SenderThread:1072 [sender.py:send_request():405] send_request: defer +2024-05-23 13:03:37,890 INFO SenderThread:1072 [sender.py:send_request_defer():609] handle sender defer: 0 +2024-05-23 13:03:37,890 INFO SenderThread:1072 [sender.py:transition_state():613] send defer: 1 +2024-05-23 13:03:37,890 DEBUG HandlerThread:1072 [handler.py:handle_request():158] handle_request: defer +2024-05-23 13:03:37,890 INFO HandlerThread:1072 [handler.py:handle_request_defer():184] handle defer: 1 +2024-05-23 13:03:37,890 DEBUG SenderThread:1072 [sender.py:send_request():405] send_request: defer +2024-05-23 13:03:37,890 INFO SenderThread:1072 [sender.py:send_request_defer():609] handle sender defer: 1 +2024-05-23 13:03:37,890 INFO SenderThread:1072 [sender.py:transition_state():613] send defer: 2 +2024-05-23 13:03:37,890 DEBUG HandlerThread:1072 [handler.py:handle_request():158] handle_request: defer +2024-05-23 13:03:37,890 INFO HandlerThread:1072 [handler.py:handle_request_defer():184] handle defer: 2 +2024-05-23 13:03:37,890 INFO HandlerThread:1072 [system_monitor.py:finish():203] Stopping system monitor +2024-05-23 13:03:37,890 DEBUG SystemMonitor:1072 [system_monitor.py:_start():172] Starting system metrics aggregation loop +2024-05-23 13:03:37,890 DEBUG SystemMonitor:1072 [system_monitor.py:_start():179] Finished system metrics aggregation loop +2024-05-23 13:03:37,891 DEBUG SystemMonitor:1072 [system_monitor.py:_start():183] Publishing last batch of metrics +2024-05-23 13:03:37,893 INFO HandlerThread:1072 [interfaces.py:finish():200] Joined cpu monitor +2024-05-23 13:03:37,893 INFO HandlerThread:1072 [interfaces.py:finish():200] Joined disk monitor +2024-05-23 13:03:37,893 INFO HandlerThread:1072 [interfaces.py:finish():200] Joined memory monitor +2024-05-23 13:03:37,894 INFO HandlerThread:1072 [interfaces.py:finish():200] Joined network monitor +2024-05-23 13:03:37,894 DEBUG SenderThread:1072 [sender.py:send_request():405] send_request: defer +2024-05-23 13:03:37,894 INFO SenderThread:1072 [sender.py:send_request_defer():609] handle sender defer: 2 +2024-05-23 13:03:37,894 INFO SenderThread:1072 [sender.py:transition_state():613] send defer: 3 +2024-05-23 13:03:37,894 DEBUG SenderThread:1072 [sender.py:send():378] send: stats +2024-05-23 13:03:37,895 DEBUG HandlerThread:1072 [handler.py:handle_request():158] handle_request: defer +2024-05-23 13:03:37,895 INFO HandlerThread:1072 [handler.py:handle_request_defer():184] handle defer: 3 +2024-05-23 13:03:37,895 DEBUG SenderThread:1072 [sender.py:send_request():405] send_request: defer +2024-05-23 13:03:37,895 INFO SenderThread:1072 [sender.py:send_request_defer():609] handle sender defer: 3 +2024-05-23 13:03:37,896 INFO SenderThread:1072 [sender.py:transition_state():613] send defer: 4 +2024-05-23 13:03:37,896 DEBUG HandlerThread:1072 [handler.py:handle_request():158] handle_request: defer +2024-05-23 13:03:37,896 INFO HandlerThread:1072 [handler.py:handle_request_defer():184] handle defer: 4 +2024-05-23 13:03:37,896 DEBUG SenderThread:1072 [sender.py:send_request():405] send_request: defer +2024-05-23 13:03:37,896 INFO SenderThread:1072 [sender.py:send_request_defer():609] handle sender defer: 4 +2024-05-23 13:03:37,896 INFO SenderThread:1072 [sender.py:transition_state():613] send defer: 5 +2024-05-23 13:03:37,896 DEBUG HandlerThread:1072 [handler.py:handle_request():158] handle_request: defer +2024-05-23 13:03:37,896 INFO HandlerThread:1072 [handler.py:handle_request_defer():184] handle defer: 5 +2024-05-23 13:03:37,896 DEBUG SenderThread:1072 [sender.py:send():378] send: summary +2024-05-23 13:03:37,897 INFO SenderThread:1072 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end +2024-05-23 13:03:37,897 DEBUG SenderThread:1072 [sender.py:send_request():405] send_request: defer +2024-05-23 13:03:37,897 INFO SenderThread:1072 [sender.py:send_request_defer():609] handle sender defer: 5 +2024-05-23 13:03:37,897 INFO SenderThread:1072 [sender.py:transition_state():613] send defer: 6 +2024-05-23 13:03:37,897 DEBUG HandlerThread:1072 [handler.py:handle_request():158] handle_request: defer +2024-05-23 13:03:37,897 INFO HandlerThread:1072 [handler.py:handle_request_defer():184] handle defer: 6 +2024-05-23 13:03:37,897 DEBUG SenderThread:1072 [sender.py:send_request():405] send_request: defer +2024-05-23 13:03:37,897 INFO SenderThread:1072 [sender.py:send_request_defer():609] handle sender defer: 6 +2024-05-23 13:03:37,902 DEBUG HandlerThread:1072 [handler.py:handle_request():158] handle_request: status_report +2024-05-23 13:03:37,997 INFO SenderThread:1072 [sender.py:transition_state():613] send defer: 7 +2024-05-23 13:03:37,997 DEBUG HandlerThread:1072 [handler.py:handle_request():158] handle_request: defer +2024-05-23 13:03:37,997 INFO HandlerThread:1072 [handler.py:handle_request_defer():184] handle defer: 7 +2024-05-23 13:03:37,997 DEBUG SenderThread:1072 [sender.py:send_request():405] send_request: defer +2024-05-23 13:03:37,998 INFO SenderThread:1072 [sender.py:send_request_defer():609] handle sender defer: 7 +2024-05-23 13:03:38,733 INFO Thread-12 :1072 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240523_130326-1oo0voi6/files/config.yaml +2024-05-23 13:03:38,734 INFO Thread-12 :1072 [dir_watcher.py:_on_file_created():271] file/dir created: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240523_130326-1oo0voi6/files/wandb-summary.json +2024-05-23 13:03:38,887 DEBUG HandlerThread:1072 [handler.py:handle_request():158] handle_request: poll_exit +2024-05-23 13:03:39,246 INFO SenderThread:1072 [sender.py:transition_state():613] send defer: 8 +2024-05-23 13:03:39,246 DEBUG SenderThread:1072 [sender.py:send_request():405] send_request: poll_exit +2024-05-23 13:03:39,246 DEBUG HandlerThread:1072 [handler.py:handle_request():158] handle_request: defer +2024-05-23 13:03:39,246 INFO HandlerThread:1072 [handler.py:handle_request_defer():184] handle defer: 8 +2024-05-23 13:03:39,246 DEBUG SenderThread:1072 [sender.py:send_request():405] send_request: defer +2024-05-23 13:03:39,246 INFO SenderThread:1072 [sender.py:send_request_defer():609] handle sender defer: 8 +2024-05-23 13:03:39,246 INFO SenderThread:1072 [job_builder.py:build():432] Attempting to build job artifact +2024-05-23 13:03:39,247 INFO SenderThread:1072 [job_builder.py:_get_source_type():576] no source found +2024-05-23 13:03:39,247 INFO SenderThread:1072 [sender.py:transition_state():613] send defer: 9 +2024-05-23 13:03:39,247 DEBUG HandlerThread:1072 [handler.py:handle_request():158] handle_request: defer +2024-05-23 13:03:39,247 INFO HandlerThread:1072 [handler.py:handle_request_defer():184] handle defer: 9 +2024-05-23 13:03:39,247 DEBUG SenderThread:1072 [sender.py:send_request():405] send_request: defer +2024-05-23 13:03:39,247 INFO SenderThread:1072 [sender.py:send_request_defer():609] handle sender defer: 9 +2024-05-23 13:03:39,247 INFO SenderThread:1072 [dir_watcher.py:finish():358] shutting down directory watcher +2024-05-23 13:03:39,735 INFO SenderThread:1072 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240523_130326-1oo0voi6/files/output.log +2024-05-23 13:03:39,735 INFO SenderThread:1072 [dir_watcher.py:finish():388] scan: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240523_130326-1oo0voi6/files +2024-05-23 13:03:39,735 INFO SenderThread:1072 [dir_watcher.py:finish():402] scan save: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240523_130326-1oo0voi6/files/wandb-summary.json wandb-summary.json +2024-05-23 13:03:39,735 INFO SenderThread:1072 [dir_watcher.py:finish():402] scan save: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240523_130326-1oo0voi6/files/wandb-metadata.json wandb-metadata.json +2024-05-23 13:03:39,736 INFO SenderThread:1072 [dir_watcher.py:finish():402] scan save: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240523_130326-1oo0voi6/files/requirements.txt requirements.txt +2024-05-23 13:03:39,738 INFO SenderThread:1072 [dir_watcher.py:finish():402] scan save: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240523_130326-1oo0voi6/files/output.log output.log +2024-05-23 13:03:39,740 INFO SenderThread:1072 [dir_watcher.py:finish():402] scan save: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240523_130326-1oo0voi6/files/config.yaml config.yaml +2024-05-23 13:03:39,742 INFO SenderThread:1072 [sender.py:transition_state():613] send defer: 10 +2024-05-23 13:03:39,743 DEBUG HandlerThread:1072 [handler.py:handle_request():158] handle_request: defer +2024-05-23 13:03:39,743 INFO HandlerThread:1072 [handler.py:handle_request_defer():184] handle defer: 10 +2024-05-23 13:03:39,743 DEBUG SenderThread:1072 [sender.py:send_request():405] send_request: defer +2024-05-23 13:03:39,745 INFO SenderThread:1072 [sender.py:send_request_defer():609] handle sender defer: 10 +2024-05-23 13:03:39,745 INFO SenderThread:1072 [file_pusher.py:finish():169] shutting down file pusher +2024-05-23 13:03:39,888 DEBUG HandlerThread:1072 [handler.py:handle_request():158] handle_request: poll_exit +2024-05-23 13:03:39,888 DEBUG SenderThread:1072 [sender.py:send_request():405] send_request: poll_exit +2024-05-23 13:03:40,056 INFO wandb-upload_0:1072 [upload_job.py:push():130] Uploaded file /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240523_130326-1oo0voi6/files/wandb-summary.json +2024-05-23 13:03:40,347 INFO wandb-upload_2:1072 [upload_job.py:push():130] Uploaded file /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240523_130326-1oo0voi6/files/output.log +2024-05-23 13:03:40,355 INFO wandb-upload_1:1072 [upload_job.py:push():130] Uploaded file /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240523_130326-1oo0voi6/files/requirements.txt +2024-05-23 13:03:40,407 INFO wandb-upload_3:1072 [upload_job.py:push():130] Uploaded file /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240523_130326-1oo0voi6/files/config.yaml +2024-05-23 13:03:40,607 INFO Thread-11 (_thread_body):1072 [sender.py:transition_state():613] send defer: 11 +2024-05-23 13:03:40,608 DEBUG HandlerThread:1072 [handler.py:handle_request():158] handle_request: defer +2024-05-23 13:03:40,608 INFO HandlerThread:1072 [handler.py:handle_request_defer():184] handle defer: 11 +2024-05-23 13:03:40,608 DEBUG SenderThread:1072 [sender.py:send_request():405] send_request: defer +2024-05-23 13:03:40,608 INFO SenderThread:1072 [sender.py:send_request_defer():609] handle sender defer: 11 +2024-05-23 13:03:40,608 INFO SenderThread:1072 [file_pusher.py:join():175] waiting for file pusher +2024-05-23 13:03:40,608 INFO SenderThread:1072 [sender.py:transition_state():613] send defer: 12 +2024-05-23 13:03:40,608 DEBUG HandlerThread:1072 [handler.py:handle_request():158] handle_request: defer +2024-05-23 13:03:40,608 INFO HandlerThread:1072 [handler.py:handle_request_defer():184] handle defer: 12 +2024-05-23 13:03:40,609 DEBUG SenderThread:1072 [sender.py:send_request():405] send_request: defer +2024-05-23 13:03:40,609 INFO SenderThread:1072 [sender.py:send_request_defer():609] handle sender defer: 12 +2024-05-23 13:03:40,609 INFO SenderThread:1072 [file_stream.py:finish():601] file stream finish called +2024-05-23 13:03:40,670 INFO SenderThread:1072 [file_stream.py:finish():605] file stream finish is done +2024-05-23 13:03:40,670 INFO SenderThread:1072 [sender.py:transition_state():613] send defer: 13 +2024-05-23 13:03:40,670 DEBUG HandlerThread:1072 [handler.py:handle_request():158] handle_request: defer +2024-05-23 13:03:40,670 INFO HandlerThread:1072 [handler.py:handle_request_defer():184] handle defer: 13 +2024-05-23 13:03:40,671 DEBUG SenderThread:1072 [sender.py:send_request():405] send_request: defer +2024-05-23 13:03:40,671 INFO SenderThread:1072 [sender.py:send_request_defer():609] handle sender defer: 13 +2024-05-23 13:03:40,671 INFO SenderThread:1072 [sender.py:transition_state():613] send defer: 14 +2024-05-23 13:03:40,671 DEBUG HandlerThread:1072 [handler.py:handle_request():158] handle_request: defer +2024-05-23 13:03:40,671 INFO HandlerThread:1072 [handler.py:handle_request_defer():184] handle defer: 14 +2024-05-23 13:03:40,671 DEBUG SenderThread:1072 [sender.py:send():378] send: final +2024-05-23 13:03:40,671 DEBUG SenderThread:1072 [sender.py:send():378] send: footer +2024-05-23 13:03:40,671 DEBUG SenderThread:1072 [sender.py:send_request():405] send_request: defer +2024-05-23 13:03:40,671 INFO SenderThread:1072 [sender.py:send_request_defer():609] handle sender defer: 14 +2024-05-23 13:03:40,672 DEBUG HandlerThread:1072 [handler.py:handle_request():158] handle_request: poll_exit +2024-05-23 13:03:40,672 DEBUG SenderThread:1072 [sender.py:send_request():405] send_request: poll_exit +2024-05-23 13:03:40,672 DEBUG HandlerThread:1072 [handler.py:handle_request():158] handle_request: poll_exit +2024-05-23 13:03:40,672 DEBUG HandlerThread:1072 [handler.py:handle_request():158] handle_request: server_info +2024-05-23 13:03:40,672 DEBUG HandlerThread:1072 [handler.py:handle_request():158] handle_request: get_summary +2024-05-23 13:03:40,672 DEBUG HandlerThread:1072 [handler.py:handle_request():158] handle_request: sampled_history +2024-05-23 13:03:40,673 DEBUG HandlerThread:1072 [handler.py:handle_request():158] handle_request: internal_messages +2024-05-23 13:03:40,673 DEBUG SenderThread:1072 [sender.py:send_request():405] send_request: poll_exit +2024-05-23 13:03:40,673 DEBUG SenderThread:1072 [sender.py:send_request():405] send_request: server_info +2024-05-23 13:03:40,727 INFO MainThread:1072 [wandb_run.py:_footer_history_summary_info():3994] rendering history +2024-05-23 13:03:40,727 INFO MainThread:1072 [wandb_run.py:_footer_history_summary_info():4026] rendering summary +2024-05-23 13:03:40,727 INFO MainThread:1072 [wandb_run.py:_footer_sync_info():3953] logging synced files +2024-05-23 13:03:40,727 DEBUG HandlerThread:1072 [handler.py:handle_request():158] handle_request: shutdown +2024-05-23 13:03:40,728 INFO HandlerThread:1072 [handler.py:finish():882] shutting down handler +2024-05-23 13:03:41,673 INFO WriterThread:1072 [datastore.py:close():296] close: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240523_130326-1oo0voi6/run-1oo0voi6.wandb +2024-05-23 13:03:41,727 INFO SenderThread:1072 [sender.py:finish():1545] shutting down sender +2024-05-23 13:03:41,727 INFO SenderThread:1072 [file_pusher.py:finish():169] shutting down file pusher +2024-05-23 13:03:41,727 INFO SenderThread:1072 [file_pusher.py:join():175] waiting for file pusher diff --git a/lm-evaluation-harness/wandb/run-20240523_130326-1oo0voi6/logs/debug.log b/lm-evaluation-harness/wandb/run-20240523_130326-1oo0voi6/logs/debug.log new file mode 100644 index 0000000000000000000000000000000000000000..6fd16030e90159e6e630f950cfc723cff954ec7d --- /dev/null +++ b/lm-evaluation-harness/wandb/run-20240523_130326-1oo0voi6/logs/debug.log @@ -0,0 +1,29 @@ +2024-05-23 13:03:26,415 INFO MainThread:917 [wandb_setup.py:_flush():76] Current SDK version is 0.17.0 +2024-05-23 13:03:26,415 INFO MainThread:917 [wandb_setup.py:_flush():76] Configure stats pid to 917 +2024-05-23 13:03:26,415 INFO MainThread:917 [wandb_setup.py:_flush():76] Loading settings from /root/.config/wandb/settings +2024-05-23 13:03:26,415 INFO MainThread:917 [wandb_setup.py:_flush():76] Loading settings from /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/settings +2024-05-23 13:03:26,415 INFO MainThread:917 [wandb_setup.py:_flush():76] Loading settings from environment variables: {} +2024-05-23 13:03:26,415 INFO MainThread:917 [wandb_setup.py:_flush():76] Applying setup settings: {'_disable_service': False} +2024-05-23 13:03:26,415 WARNING MainThread:917 [wandb_setup.py:_flush():76] Could not find program at -m lm_eval.__main__ +2024-05-23 13:03:26,415 INFO MainThread:917 [wandb_setup.py:_flush():76] Inferring run settings from compute environment: {'program_relpath': None, 'program': '-m lm_eval.__main__'} +2024-05-23 13:03:26,416 INFO MainThread:917 [wandb_setup.py:_flush():76] Applying login settings: {} +2024-05-23 13:03:26,416 INFO MainThread:917 [wandb_init.py:_log_setup():520] Logging user logs to /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240523_130326-1oo0voi6/logs/debug.log +2024-05-23 13:03:26,416 INFO MainThread:917 [wandb_init.py:_log_setup():521] Logging internal logs to /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240523_130326-1oo0voi6/logs/debug-internal.log +2024-05-23 13:03:26,416 INFO MainThread:917 [wandb_init.py:init():560] calling init triggers +2024-05-23 13:03:26,416 INFO MainThread:917 [wandb_init.py:init():567] wandb.init called with sweep_config: {} +config: {} +2024-05-23 13:03:26,416 INFO MainThread:917 [wandb_init.py:init():610] starting backend +2024-05-23 13:03:26,416 INFO MainThread:917 [wandb_init.py:init():614] setting up manager +2024-05-23 13:03:26,419 INFO MainThread:917 [backend.py:_multiprocessing_setup():105] multiprocessing start_methods=fork,spawn,forkserver, using: spawn +2024-05-23 13:03:26,420 INFO MainThread:917 [wandb_init.py:init():622] backend started and connected +2024-05-23 13:03:26,424 INFO MainThread:917 [wandb_init.py:init():711] updated telemetry +2024-05-23 13:03:26,432 INFO MainThread:917 [wandb_init.py:init():744] communicating run to backend with 90.0 second timeout +2024-05-23 13:03:26,726 INFO MainThread:917 [wandb_run.py:_on_init():2396] communicating current version +2024-05-23 13:03:26,843 INFO MainThread:917 [wandb_run.py:_on_init():2405] got version response +2024-05-23 13:03:26,844 INFO MainThread:917 [wandb_init.py:init():795] starting run threads in backend +2024-05-23 13:03:27,126 INFO MainThread:917 [wandb_run.py:_console_start():2374] atexit reg +2024-05-23 13:03:27,126 INFO MainThread:917 [wandb_run.py:_redirect():2229] redirect: wrap_raw +2024-05-23 13:03:27,126 INFO MainThread:917 [wandb_run.py:_redirect():2294] Wrapping output streams. +2024-05-23 13:03:27,127 INFO MainThread:917 [wandb_run.py:_redirect():2319] Redirects installed. +2024-05-23 13:03:27,129 INFO MainThread:917 [wandb_init.py:init():838] run started, returning control to user process +2024-05-23 13:03:41,728 WARNING MsgRouterThr:917 [router.py:message_loop():77] message_loop has been closed diff --git a/lm-evaluation-harness/wandb/run-20240523_130326-1oo0voi6/run-1oo0voi6.wandb b/lm-evaluation-harness/wandb/run-20240523_130326-1oo0voi6/run-1oo0voi6.wandb new file mode 100644 index 0000000000000000000000000000000000000000..502d015533f03c4721f653240157255eb8dc158f Binary files /dev/null and b/lm-evaluation-harness/wandb/run-20240523_130326-1oo0voi6/run-1oo0voi6.wandb differ diff --git a/lm-evaluation-harness/wandb/run-20240523_130448-vm5e7ag8/files/config.yaml b/lm-evaluation-harness/wandb/run-20240523_130448-vm5e7ag8/files/config.yaml new file mode 100644 index 0000000000000000000000000000000000000000..699058f9663d3a67da9aee10d155e77b52da6c83 --- /dev/null +++ b/lm-evaluation-harness/wandb/run-20240523_130448-vm5e7ag8/files/config.yaml @@ -0,0 +1,43 @@ +wandb_version: 1 + +_wandb: + desc: null + value: + python_version: 3.10.12 + cli_version: 0.17.0 + framework: huggingface + huggingface_version: 4.41.1 + is_jupyter_run: false + is_kaggle_kernel: false + start_time: 1716469488 + t: + 1: + - 1 + - 5 + - 11 + - 49 + - 51 + - 53 + - 55 + - 71 + - 98 + - 100 + 2: + - 1 + - 5 + - 11 + - 49 + - 51 + - 53 + - 55 + - 71 + - 98 + - 100 + 3: + - 23 + 4: 3.10.12 + 5: 0.17.0 + 6: 4.41.1 + 8: + - 5 + 13: linux-x86_64 diff --git a/lm-evaluation-harness/wandb/run-20240523_130448-vm5e7ag8/files/output.log b/lm-evaluation-harness/wandb/run-20240523_130448-vm5e7ag8/files/output.log new file mode 100644 index 0000000000000000000000000000000000000000..f239d543b41c5ccfdb20f8e5b4e3da336ba8feff --- /dev/null +++ b/lm-evaluation-harness/wandb/run-20240523_130448-vm5e7ag8/files/output.log @@ -0,0 +1,34 @@ + +2024-05-23:13:04:48,983 INFO [__main__.py:251] Verbosity set to INFO +2024-05-23:13:04:57,475 INFO [__main__.py:335] Selected Tasks: ['arc_easy', 'hellaswag', 'mrpc', 'openbookqa', 'sst2', 'winogrande'] +2024-05-23:13:04:57,476 INFO [evaluator.py:131] Setting random seed to 0 | Setting numpy seed to 1234 | Setting torch manual seed to 1234 +2024-05-23:13:04:57,477 INFO [evaluator.py:177] Initializing hf model, with arguments: {'pretrained': '/mnt/weka/peacock/experiments/llama/checkpoint/llamav2-3b//hf_ckpt//global_step14000'} +2024-05-23:13:04:59,760 INFO [huggingface.py:164] Using device 'cuda' +Traceback (most recent call last): + File "/usr/lib/python3.10/runpy.py", line 196, in _run_module_as_main + return _run_code(code, main_globals, None, + File "/usr/lib/python3.10/runpy.py", line 86, in _run_code + exec(code, run_globals) + File "/mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/lm_eval/__main__.py", line 417, in + cli_evaluate() + File "/mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/lm_eval/__main__.py", line 341, in cli_evaluate + results = evaluator.simple_evaluate( + File "/mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/lm_eval/utils.py", line 288, in _wrapper + return fn(*args, **kwargs) + File "/mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/lm_eval/evaluator.py", line 180, in simple_evaluate + lm = lm_eval.api.registry.get_model(model).create_from_arg_string( + File "/mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/lm_eval/api/model.py", line 134, in create_from_arg_string + return cls(**args, **args2) + File "/mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/lm_eval/models/huggingface.py", line 190, in __init__ + self._get_config( + File "/mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/lm_eval/models/huggingface.py", line 471, in _get_config + self._config = transformers.AutoConfig.from_pretrained( + File "/usr/local/lib/python3.10/dist-packages/transformers/models/auto/configuration_auto.py", line 934, in from_pretrained + config_dict, unused_kwargs = PretrainedConfig.get_config_dict(pretrained_model_name_or_path, **kwargs) + File "/usr/local/lib/python3.10/dist-packages/transformers/configuration_utils.py", line 632, in get_config_dict + config_dict, kwargs = cls._get_config_dict(pretrained_model_name_or_path, **kwargs) + File "/usr/local/lib/python3.10/dist-packages/transformers/configuration_utils.py", line 689, in _get_config_dict + resolved_config_file = cached_file( + File "/usr/local/lib/python3.10/dist-packages/transformers/utils/hub.py", line 370, in cached_file + raise EnvironmentError( +OSError: /mnt/weka/peacock/experiments/llama/checkpoint/llamav2-3b//hf_ckpt//global_step14000 does not appear to have a file named config.json. Checkout 'https://huggingface.co//mnt/weka/peacock/experiments/llama/checkpoint/llamav2-3b//hf_ckpt//global_step14000/tree/main' for available files. \ No newline at end of file diff --git a/lm-evaluation-harness/wandb/run-20240523_130448-vm5e7ag8/files/requirements.txt b/lm-evaluation-harness/wandb/run-20240523_130448-vm5e7ag8/files/requirements.txt new file mode 100644 index 0000000000000000000000000000000000000000..f675c3016b5332c1acf28f436e0b60adeead9c12 --- /dev/null +++ b/lm-evaluation-harness/wandb/run-20240523_130448-vm5e7ag8/files/requirements.txt @@ -0,0 +1,155 @@ +DataProperty==1.0.1 +GitPython==3.1.43 +Jinja2==3.1.4 +Markdown==3.6 +MarkupSafe==2.1.5 +Pillow-SIMD==7.0.0.post3 +PyYAML==6.0 +Werkzeug==3.0.3 +absl-py==2.1.0 +accelerate==0.30.1 +aiohttp==3.9.5 +aiosignal==1.3.1 +async-timeout==4.0.3 +attrs==23.2.0 +av==9.2.0 +cachetools==5.3.3 +certifi==2024.2.2 +cffi==1.15.1 +cfgv==3.4.0 +chardet==5.2.0 +charset-normalizer==3.3.2 +click==8.1.7 +cmake==3.29.2 +colorama==0.4.6 +datasets==2.19.1 +deepspeed==0.12.4+hpu.synapse.v1.15.1 +dill==0.3.8 +distlib==0.3.8 +docker-pycreds==0.4.0 +einops==0.8.0 +evaluate==0.4.2 +exceptiongroup==1.2.1 +expecttest==0.2.1 +filelock==3.14.0 +frozenlist==1.4.1 +fsspec==2024.3.1 +gitdb==4.0.11 +google-auth-oauthlib==0.4.6 +google-auth==2.29.0 +grpcio==1.63.0 +habana-media-loader==1.15.1.15 +habana-pyhlml==1.15.1.15 +habana-torch-dataloader==1.15.1.15 +habana-torch-plugin==1.15.1.15 +habana_gpu_migration==1.15.1.15 +habana_quantization_toolkit==1.15.1.15 +hjson==3.1.0 +huggingface-hub==0.23.1 +identify==2.5.36 +idna==3.7 +iniconfig==2.0.0 +joblib==1.4.2 +jsonlines==4.0.0 +lightning-habana==1.4.0 +lightning-utilities==0.11.2 +lightning==2.2.0.post0 +lm_eval==0.4.2 +lm_eval==0.4.2 +lm_eval==0.4.2 +lxml==5.2.2 +mbstrdecoder==1.1.3 +more-itertools==10.2.0 +mpi4py==3.1.4 +mpmath==1.3.0 +multidict==6.0.5 +multiprocess==0.70.16 +networkx==3.3 +ninja==1.11.1.1 +nltk==3.8.1 +nodeenv==1.8.0 +numexpr==2.10.0 +numpy==1.23.5 +oauthlib==3.2.2 +packaging==24.0 +pandas==2.0.1 +pathspec==0.12.1 +pathvalidate==3.2.0 +peft==0.11.1 +perfetto==0.7.0 +pillow==10.3.0 +pip==22.0.2 +pip==23.3.1 +platformdirs==4.2.1 +pluggy==1.5.0 +portalocker==2.8.2 +pre-commit==3.3.3 +pretty-errors==1.2.25 +protobuf==3.20.3 +psutil==5.9.8 +py-cpuinfo==9.0.0 +pyarrow-hotfix==0.6 +pyarrow==16.1.0 +pyasn1==0.6.0 +pyasn1_modules==0.4.0 +pybind11==2.10.4 +pycparser==2.22 +pydantic==1.10.13 +pynvml==8.0.4 +pytablewriter==1.2.0 +pytest==8.2.0 +python-dateutil==2.9.0.post0 +pytorch-lightning==2.2.4 +pytz==2024.1 +regex==2023.5.5 +requests-oauthlib==2.0.0 +requests==2.31.0 +rouge_score==0.1.2 +rsa==4.9 +sacrebleu==2.4.2 +safetensors==0.4.3 +scikit-learn==1.5.0 +scipy==1.13.1 +sentencepiece==0.2.0 +sentry-sdk==2.3.0 +setproctitle==1.3.3 +setuptools==59.6.0 +setuptools==69.5.1 +six==1.16.0 +smmap==5.0.1 +sqlitedict==2.1.0 +symengine==0.11.0 +sympy==1.12 +tabledata==1.3.3 +tabulate==0.9.0 +tcolorpy==0.1.6 +tdqm==0.0.1 +tensorboard-data-server==0.6.1 +tensorboard-plugin-wit==1.8.1 +tensorboard==2.11.2 +threadpoolctl==3.5.0 +tokenizers==0.19.1 +tomli==2.0.1 +torch==2.2.0a0+git8964477 +torch_tb_profiler==0.4.0 +torchaudio==2.2.0+08901ad +torchdata==0.7.1+5e6f7b7 +torchmetrics==1.4.0 +torchtext==0.17.0+400da5c +torchvision==0.17.0+b2383d4 +tqdm-multiprocess==0.0.11 +tqdm==4.66.4 +transformers==4.41.1 +typepy==1.3.2 +typing_extensions==4.11.0 +tzdata==2024.1 +urllib3==1.26.18 +virtualenv==20.26.1 +wandb==0.17.0 +wheel==0.37.1 +wheel==0.43.0 +word2number==1.1 +xxhash==3.4.1 +yamllint==1.35.1 +yarl==1.9.4 +zstandard==0.22.0 \ No newline at end of file diff --git a/lm-evaluation-harness/wandb/run-20240523_130448-vm5e7ag8/files/wandb-metadata.json b/lm-evaluation-harness/wandb/run-20240523_130448-vm5e7ag8/files/wandb-metadata.json new file mode 100644 index 0000000000000000000000000000000000000000..aa47628ab0d800e57ef4f1d0c798b3534a01fda3 --- /dev/null +++ b/lm-evaluation-harness/wandb/run-20240523_130448-vm5e7ag8/files/wandb-metadata.json @@ -0,0 +1,850 @@ +{ + "os": "Linux-5.15.0-92-generic-x86_64-with-glibc2.35", + "python": "3.10.12", + "heartbeatAt": "2024-05-23T13:04:48.781263", + "startedAt": "2024-05-23T13:04:48.261491", + "docker": null, + "cuda": null, + "args": [ + "--model", + "hf", + "--model_args", + "pretrained=/mnt/weka/peacock/experiments/llama/checkpoint/llamav2-3b//hf_ckpt//global_step14000", + "--tasks", + "hellaswag,arc_easy,openbookqa,winogrande,sst2,mrpc", + "--batch_size", + "auto", + "--wandb_args", + "project=bharatgpt,group=trial_expt_2" + ], + "state": "running", + "program": "-m lm_eval.__main__", + "codePathLocal": null, + "git": { + "remote": "https://github.com/EleutherAI/lm-evaluation-harness", + "commit": null + }, + "email": null, + "root": "/mnt/weka/peacock/idc/cronscript/lm-evaluation-harness", + "host": "peacock-evaluation-worker-0", + "username": "root", + "executable": "/usr/bin/python3", + "cpu_count": 80, + "cpu_count_logical": 160, + "cpu_freq": { + "current": 2365.89415, + "min": 800.0, + "max": 3400.0 + }, + "cpu_freq_per_core": [ + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + } + ], + "disk": { + "/": { + "total": 877.6341285705566, + "used": 211.62759399414062 + } + }, + "memory": { + "total": 1007.4379539489746 + } +} diff --git a/lm-evaluation-harness/wandb/run-20240523_130448-vm5e7ag8/files/wandb-summary.json b/lm-evaluation-harness/wandb/run-20240523_130448-vm5e7ag8/files/wandb-summary.json new file mode 100644 index 0000000000000000000000000000000000000000..8bf99d152ad35c3699ec8600ecb8b169d4e35875 --- /dev/null +++ b/lm-evaluation-harness/wandb/run-20240523_130448-vm5e7ag8/files/wandb-summary.json @@ -0,0 +1 @@ +{"_wandb": {"runtime": 11}} \ No newline at end of file diff --git a/lm-evaluation-harness/wandb/run-20240523_130448-vm5e7ag8/logs/debug-internal.log b/lm-evaluation-harness/wandb/run-20240523_130448-vm5e7ag8/logs/debug-internal.log new file mode 100644 index 0000000000000000000000000000000000000000..98d530934d06bd2bb9bb49a5fcd2216babb6ab38 --- /dev/null +++ b/lm-evaluation-harness/wandb/run-20240523_130448-vm5e7ag8/logs/debug-internal.log @@ -0,0 +1,183 @@ +2024-05-23 13:04:48,284 INFO StreamThr :1586 [internal.py:wandb_internal():85] W&B internal server running at pid: 1586, started at: 2024-05-23 13:04:48.281109 +2024-05-23 13:04:48,288 DEBUG HandlerThread:1586 [handler.py:handle_request():158] handle_request: status +2024-05-23 13:04:48,288 INFO WriterThread:1586 [datastore.py:open_for_write():87] open: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240523_130448-vm5e7ag8/run-vm5e7ag8.wandb +2024-05-23 13:04:48,293 DEBUG SenderThread:1586 [sender.py:send():378] send: header +2024-05-23 13:04:48,294 DEBUG SenderThread:1586 [sender.py:send():378] send: run +2024-05-23 13:04:48,541 INFO SenderThread:1586 [dir_watcher.py:__init__():211] watching files in: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240523_130448-vm5e7ag8/files +2024-05-23 13:04:48,541 INFO SenderThread:1586 [sender.py:_start_run_threads():1123] run started: vm5e7ag8 with start time 1716469488.280964 +2024-05-23 13:04:48,542 DEBUG HandlerThread:1586 [handler.py:handle_request():158] handle_request: check_version +2024-05-23 13:04:48,542 DEBUG SenderThread:1586 [sender.py:send_request():405] send_request: check_version +2024-05-23 13:04:48,704 DEBUG HandlerThread:1586 [handler.py:handle_request():158] handle_request: run_start +2024-05-23 13:04:48,706 DEBUG HandlerThread:1586 [system_info.py:__init__():26] System info init +2024-05-23 13:04:48,706 DEBUG HandlerThread:1586 [system_info.py:__init__():41] System info init done +2024-05-23 13:04:48,706 INFO HandlerThread:1586 [system_monitor.py:start():194] Starting system monitor +2024-05-23 13:04:48,706 INFO SystemMonitor:1586 [system_monitor.py:_start():158] Starting system asset monitoring threads +2024-05-23 13:04:48,706 INFO HandlerThread:1586 [system_monitor.py:probe():214] Collecting system info +2024-05-23 13:04:48,713 INFO SystemMonitor:1586 [interfaces.py:start():188] Started cpu monitoring +2024-05-23 13:04:48,719 INFO SystemMonitor:1586 [interfaces.py:start():188] Started disk monitoring +2024-05-23 13:04:48,719 INFO SystemMonitor:1586 [interfaces.py:start():188] Started memory monitoring +2024-05-23 13:04:48,719 INFO SystemMonitor:1586 [interfaces.py:start():188] Started network monitoring +2024-05-23 13:04:48,781 DEBUG HandlerThread:1586 [system_info.py:probe():150] Probing system +2024-05-23 13:04:48,784 DEBUG HandlerThread:1586 [system_info.py:_probe_git():135] Probing git +2024-05-23 13:04:48,794 ERROR HandlerThread:1586 [gitlib.py:root():92] git root error: Cmd('git') failed due to: exit code(128) + cmdline: git rev-parse --show-toplevel + stderr: 'fatal: detected dubious ownership in repository at '/mnt/weka/peacock/idc/cronscript/lm-evaluation-harness' +To add an exception for this directory, call: + + git config --global --add safe.directory /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness' +2024-05-23 13:04:48,794 DEBUG HandlerThread:1586 [system_info.py:_probe_git():143] Probing git done +2024-05-23 13:04:48,794 DEBUG HandlerThread:1586 [system_info.py:probe():198] Probing system done +2024-05-23 13:04:48,794 DEBUG HandlerThread:1586 [system_monitor.py:probe():223] {'os': 'Linux-5.15.0-92-generic-x86_64-with-glibc2.35', 'python': '3.10.12', 'heartbeatAt': '2024-05-23T13:04:48.781263', 'startedAt': '2024-05-23T13:04:48.261491', 'docker': None, 'cuda': None, 'args': ('--model', 'hf', '--model_args', 'pretrained=/mnt/weka/peacock/experiments/llama/checkpoint/llamav2-3b//hf_ckpt//global_step14000', '--tasks', 'hellaswag,arc_easy,openbookqa,winogrande,sst2,mrpc', '--batch_size', 'auto', '--wandb_args', 'project=bharatgpt,group=trial_expt_2'), 'state': 'running', 'program': '-m lm_eval.__main__', 'codePathLocal': None, 'git': {'remote': 'https://github.com/EleutherAI/lm-evaluation-harness', 'commit': None}, 'email': None, 'root': '/mnt/weka/peacock/idc/cronscript/lm-evaluation-harness', 'host': 'peacock-evaluation-worker-0', 'username': 'root', 'executable': '/usr/bin/python3', 'cpu_count': 80, 'cpu_count_logical': 160, 'cpu_freq': {'current': 2365.89415, 'min': 800.0, 'max': 3400.0}, 'cpu_freq_per_core': [{'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}], 'disk': {'/': {'total': 877.6341285705566, 'used': 211.62759399414062}}, 'memory': {'total': 1007.4379539489746}} +2024-05-23 13:04:48,794 INFO HandlerThread:1586 [system_monitor.py:probe():224] Finished collecting system info +2024-05-23 13:04:48,794 INFO HandlerThread:1586 [system_monitor.py:probe():227] Publishing system info +2024-05-23 13:04:48,797 INFO HandlerThread:1586 [system_monitor.py:probe():229] Finished publishing system info +2024-05-23 13:04:48,802 DEBUG SenderThread:1586 [sender.py:send():378] send: files +2024-05-23 13:04:48,803 INFO SenderThread:1586 [sender.py:_save_file():1389] saving file wandb-metadata.json with policy now +2024-05-23 13:04:48,976 DEBUG HandlerThread:1586 [handler.py:handle_request():158] handle_request: python_packages +2024-05-23 13:04:48,976 DEBUG SenderThread:1586 [sender.py:send_request():405] send_request: python_packages +2024-05-23 13:04:48,977 DEBUG HandlerThread:1586 [handler.py:handle_request():158] handle_request: stop_status +2024-05-23 13:04:48,978 DEBUG SenderThread:1586 [sender.py:send_request():405] send_request: stop_status +2024-05-23 13:04:49,135 DEBUG SenderThread:1586 [sender.py:send():378] send: telemetry +2024-05-23 13:04:49,355 INFO wandb-upload_0:1586 [upload_job.py:push():130] Uploaded file /tmp/tmpj2js6yoqwandb/bh6r6bog-wandb-metadata.json +2024-05-23 13:04:49,543 INFO Thread-12 :1586 [dir_watcher.py:_on_file_created():271] file/dir created: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240523_130448-vm5e7ag8/files/requirements.txt +2024-05-23 13:04:49,543 INFO Thread-12 :1586 [dir_watcher.py:_on_file_created():271] file/dir created: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240523_130448-vm5e7ag8/files/output.log +2024-05-23 13:04:49,543 INFO Thread-12 :1586 [dir_watcher.py:_on_file_created():271] file/dir created: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240523_130448-vm5e7ag8/files/wandb-metadata.json +2024-05-23 13:04:51,542 INFO Thread-12 :1586 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240523_130448-vm5e7ag8/files/output.log +2024-05-23 13:04:54,140 DEBUG HandlerThread:1586 [handler.py:handle_request():158] handle_request: status_report +2024-05-23 13:04:59,478 DEBUG HandlerThread:1586 [handler.py:handle_request():158] handle_request: status_report +2024-05-23 13:04:59,551 INFO Thread-12 :1586 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240523_130448-vm5e7ag8/files/output.log +2024-05-23 13:04:59,771 DEBUG SenderThread:1586 [sender.py:send():378] send: exit +2024-05-23 13:04:59,771 INFO SenderThread:1586 [sender.py:send_exit():585] handling exit code: 1 +2024-05-23 13:04:59,772 INFO SenderThread:1586 [sender.py:send_exit():587] handling runtime: 11 +2024-05-23 13:04:59,773 INFO SenderThread:1586 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end +2024-05-23 13:04:59,773 INFO SenderThread:1586 [sender.py:send_exit():593] send defer +2024-05-23 13:04:59,773 DEBUG HandlerThread:1586 [handler.py:handle_request():158] handle_request: defer +2024-05-23 13:04:59,774 INFO HandlerThread:1586 [handler.py:handle_request_defer():184] handle defer: 0 +2024-05-23 13:04:59,774 DEBUG SenderThread:1586 [sender.py:send_request():405] send_request: defer +2024-05-23 13:04:59,774 INFO SenderThread:1586 [sender.py:send_request_defer():609] handle sender defer: 0 +2024-05-23 13:04:59,774 INFO SenderThread:1586 [sender.py:transition_state():613] send defer: 1 +2024-05-23 13:04:59,774 DEBUG HandlerThread:1586 [handler.py:handle_request():158] handle_request: defer +2024-05-23 13:04:59,774 INFO HandlerThread:1586 [handler.py:handle_request_defer():184] handle defer: 1 +2024-05-23 13:04:59,774 DEBUG SenderThread:1586 [sender.py:send_request():405] send_request: defer +2024-05-23 13:04:59,774 INFO SenderThread:1586 [sender.py:send_request_defer():609] handle sender defer: 1 +2024-05-23 13:04:59,774 INFO SenderThread:1586 [sender.py:transition_state():613] send defer: 2 +2024-05-23 13:04:59,774 DEBUG HandlerThread:1586 [handler.py:handle_request():158] handle_request: defer +2024-05-23 13:04:59,774 INFO HandlerThread:1586 [handler.py:handle_request_defer():184] handle defer: 2 +2024-05-23 13:04:59,774 INFO HandlerThread:1586 [system_monitor.py:finish():203] Stopping system monitor +2024-05-23 13:04:59,775 DEBUG SystemMonitor:1586 [system_monitor.py:_start():172] Starting system metrics aggregation loop +2024-05-23 13:04:59,775 DEBUG SystemMonitor:1586 [system_monitor.py:_start():179] Finished system metrics aggregation loop +2024-05-23 13:04:59,775 DEBUG SystemMonitor:1586 [system_monitor.py:_start():183] Publishing last batch of metrics +2024-05-23 13:04:59,777 INFO HandlerThread:1586 [interfaces.py:finish():200] Joined cpu monitor +2024-05-23 13:04:59,778 INFO HandlerThread:1586 [interfaces.py:finish():200] Joined disk monitor +2024-05-23 13:04:59,778 INFO HandlerThread:1586 [interfaces.py:finish():200] Joined memory monitor +2024-05-23 13:04:59,778 INFO HandlerThread:1586 [interfaces.py:finish():200] Joined network monitor +2024-05-23 13:04:59,778 DEBUG SenderThread:1586 [sender.py:send_request():405] send_request: defer +2024-05-23 13:04:59,778 INFO SenderThread:1586 [sender.py:send_request_defer():609] handle sender defer: 2 +2024-05-23 13:04:59,778 INFO SenderThread:1586 [sender.py:transition_state():613] send defer: 3 +2024-05-23 13:04:59,778 DEBUG HandlerThread:1586 [handler.py:handle_request():158] handle_request: defer +2024-05-23 13:04:59,779 INFO HandlerThread:1586 [handler.py:handle_request_defer():184] handle defer: 3 +2024-05-23 13:04:59,779 DEBUG SenderThread:1586 [sender.py:send():378] send: stats +2024-05-23 13:04:59,780 DEBUG SenderThread:1586 [sender.py:send_request():405] send_request: defer +2024-05-23 13:04:59,780 INFO SenderThread:1586 [sender.py:send_request_defer():609] handle sender defer: 3 +2024-05-23 13:04:59,780 INFO SenderThread:1586 [sender.py:transition_state():613] send defer: 4 +2024-05-23 13:04:59,780 DEBUG HandlerThread:1586 [handler.py:handle_request():158] handle_request: defer +2024-05-23 13:04:59,780 INFO HandlerThread:1586 [handler.py:handle_request_defer():184] handle defer: 4 +2024-05-23 13:04:59,780 DEBUG SenderThread:1586 [sender.py:send_request():405] send_request: defer +2024-05-23 13:04:59,780 INFO SenderThread:1586 [sender.py:send_request_defer():609] handle sender defer: 4 +2024-05-23 13:04:59,780 INFO SenderThread:1586 [sender.py:transition_state():613] send defer: 5 +2024-05-23 13:04:59,780 DEBUG HandlerThread:1586 [handler.py:handle_request():158] handle_request: defer +2024-05-23 13:04:59,780 INFO HandlerThread:1586 [handler.py:handle_request_defer():184] handle defer: 5 +2024-05-23 13:04:59,780 DEBUG SenderThread:1586 [sender.py:send():378] send: summary +2024-05-23 13:04:59,781 INFO SenderThread:1586 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end +2024-05-23 13:04:59,781 DEBUG SenderThread:1586 [sender.py:send_request():405] send_request: defer +2024-05-23 13:04:59,781 INFO SenderThread:1586 [sender.py:send_request_defer():609] handle sender defer: 5 +2024-05-23 13:04:59,781 INFO SenderThread:1586 [sender.py:transition_state():613] send defer: 6 +2024-05-23 13:04:59,781 DEBUG HandlerThread:1586 [handler.py:handle_request():158] handle_request: defer +2024-05-23 13:04:59,782 INFO HandlerThread:1586 [handler.py:handle_request_defer():184] handle defer: 6 +2024-05-23 13:04:59,782 DEBUG SenderThread:1586 [sender.py:send_request():405] send_request: defer +2024-05-23 13:04:59,782 INFO SenderThread:1586 [sender.py:send_request_defer():609] handle sender defer: 6 +2024-05-23 13:04:59,786 DEBUG HandlerThread:1586 [handler.py:handle_request():158] handle_request: status_report +2024-05-23 13:04:59,874 INFO SenderThread:1586 [sender.py:transition_state():613] send defer: 7 +2024-05-23 13:04:59,874 DEBUG HandlerThread:1586 [handler.py:handle_request():158] handle_request: defer +2024-05-23 13:04:59,874 INFO HandlerThread:1586 [handler.py:handle_request_defer():184] handle defer: 7 +2024-05-23 13:04:59,874 DEBUG SenderThread:1586 [sender.py:send_request():405] send_request: defer +2024-05-23 13:04:59,874 INFO SenderThread:1586 [sender.py:send_request_defer():609] handle sender defer: 7 +2024-05-23 13:05:00,552 INFO Thread-12 :1586 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240523_130448-vm5e7ag8/files/config.yaml +2024-05-23 13:05:00,552 INFO Thread-12 :1586 [dir_watcher.py:_on_file_created():271] file/dir created: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240523_130448-vm5e7ag8/files/wandb-summary.json +2024-05-23 13:05:00,772 DEBUG HandlerThread:1586 [handler.py:handle_request():158] handle_request: poll_exit +2024-05-23 13:05:01,161 INFO SenderThread:1586 [sender.py:transition_state():613] send defer: 8 +2024-05-23 13:05:01,161 DEBUG SenderThread:1586 [sender.py:send_request():405] send_request: poll_exit +2024-05-23 13:05:01,161 DEBUG HandlerThread:1586 [handler.py:handle_request():158] handle_request: defer +2024-05-23 13:05:01,161 INFO HandlerThread:1586 [handler.py:handle_request_defer():184] handle defer: 8 +2024-05-23 13:05:01,161 DEBUG SenderThread:1586 [sender.py:send_request():405] send_request: defer +2024-05-23 13:05:01,161 INFO SenderThread:1586 [sender.py:send_request_defer():609] handle sender defer: 8 +2024-05-23 13:05:01,161 INFO SenderThread:1586 [job_builder.py:build():432] Attempting to build job artifact +2024-05-23 13:05:01,162 INFO SenderThread:1586 [job_builder.py:_get_source_type():576] no source found +2024-05-23 13:05:01,162 INFO SenderThread:1586 [sender.py:transition_state():613] send defer: 9 +2024-05-23 13:05:01,162 DEBUG HandlerThread:1586 [handler.py:handle_request():158] handle_request: defer +2024-05-23 13:05:01,162 INFO HandlerThread:1586 [handler.py:handle_request_defer():184] handle defer: 9 +2024-05-23 13:05:01,162 DEBUG SenderThread:1586 [sender.py:send_request():405] send_request: defer +2024-05-23 13:05:01,162 INFO SenderThread:1586 [sender.py:send_request_defer():609] handle sender defer: 9 +2024-05-23 13:05:01,162 INFO SenderThread:1586 [dir_watcher.py:finish():358] shutting down directory watcher +2024-05-23 13:05:01,554 INFO SenderThread:1586 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240523_130448-vm5e7ag8/files/output.log +2024-05-23 13:05:01,554 INFO SenderThread:1586 [dir_watcher.py:finish():388] scan: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240523_130448-vm5e7ag8/files +2024-05-23 13:05:01,556 INFO SenderThread:1586 [dir_watcher.py:finish():402] scan save: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240523_130448-vm5e7ag8/files/wandb-metadata.json wandb-metadata.json +2024-05-23 13:05:01,556 INFO SenderThread:1586 [dir_watcher.py:finish():402] scan save: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240523_130448-vm5e7ag8/files/output.log output.log +2024-05-23 13:05:01,556 INFO SenderThread:1586 [dir_watcher.py:finish():402] scan save: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240523_130448-vm5e7ag8/files/wandb-summary.json wandb-summary.json +2024-05-23 13:05:01,559 INFO SenderThread:1586 [dir_watcher.py:finish():402] scan save: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240523_130448-vm5e7ag8/files/config.yaml config.yaml +2024-05-23 13:05:01,561 INFO SenderThread:1586 [dir_watcher.py:finish():402] scan save: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240523_130448-vm5e7ag8/files/requirements.txt requirements.txt +2024-05-23 13:05:01,563 INFO SenderThread:1586 [sender.py:transition_state():613] send defer: 10 +2024-05-23 13:05:01,563 DEBUG HandlerThread:1586 [handler.py:handle_request():158] handle_request: defer +2024-05-23 13:05:01,564 INFO HandlerThread:1586 [handler.py:handle_request_defer():184] handle defer: 10 +2024-05-23 13:05:01,564 DEBUG SenderThread:1586 [sender.py:send_request():405] send_request: defer +2024-05-23 13:05:01,566 INFO SenderThread:1586 [sender.py:send_request_defer():609] handle sender defer: 10 +2024-05-23 13:05:01,566 INFO SenderThread:1586 [file_pusher.py:finish():169] shutting down file pusher +2024-05-23 13:05:01,772 DEBUG HandlerThread:1586 [handler.py:handle_request():158] handle_request: poll_exit +2024-05-23 13:05:01,772 DEBUG SenderThread:1586 [sender.py:send_request():405] send_request: poll_exit +2024-05-23 13:05:01,819 INFO wandb-upload_0:1586 [upload_job.py:push():130] Uploaded file /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240523_130448-vm5e7ag8/files/output.log +2024-05-23 13:05:02,170 INFO wandb-upload_2:1586 [upload_job.py:push():130] Uploaded file /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240523_130448-vm5e7ag8/files/config.yaml +2024-05-23 13:05:02,209 INFO wandb-upload_3:1586 [upload_job.py:push():130] Uploaded file /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240523_130448-vm5e7ag8/files/requirements.txt +2024-05-23 13:05:02,275 INFO wandb-upload_1:1586 [upload_job.py:push():130] Uploaded file /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240523_130448-vm5e7ag8/files/wandb-summary.json +2024-05-23 13:05:02,475 INFO Thread-11 (_thread_body):1586 [sender.py:transition_state():613] send defer: 11 +2024-05-23 13:05:02,475 DEBUG HandlerThread:1586 [handler.py:handle_request():158] handle_request: defer +2024-05-23 13:05:02,475 INFO HandlerThread:1586 [handler.py:handle_request_defer():184] handle defer: 11 +2024-05-23 13:05:02,475 DEBUG SenderThread:1586 [sender.py:send_request():405] send_request: defer +2024-05-23 13:05:02,475 INFO SenderThread:1586 [sender.py:send_request_defer():609] handle sender defer: 11 +2024-05-23 13:05:02,476 INFO SenderThread:1586 [file_pusher.py:join():175] waiting for file pusher +2024-05-23 13:05:02,476 INFO SenderThread:1586 [sender.py:transition_state():613] send defer: 12 +2024-05-23 13:05:02,476 DEBUG HandlerThread:1586 [handler.py:handle_request():158] handle_request: defer +2024-05-23 13:05:02,476 INFO HandlerThread:1586 [handler.py:handle_request_defer():184] handle defer: 12 +2024-05-23 13:05:02,476 DEBUG SenderThread:1586 [sender.py:send_request():405] send_request: defer +2024-05-23 13:05:02,476 INFO SenderThread:1586 [sender.py:send_request_defer():609] handle sender defer: 12 +2024-05-23 13:05:02,476 INFO SenderThread:1586 [file_stream.py:finish():601] file stream finish called +2024-05-23 13:05:02,535 INFO SenderThread:1586 [file_stream.py:finish():605] file stream finish is done +2024-05-23 13:05:02,535 INFO SenderThread:1586 [sender.py:transition_state():613] send defer: 13 +2024-05-23 13:05:02,535 DEBUG HandlerThread:1586 [handler.py:handle_request():158] handle_request: defer +2024-05-23 13:05:02,535 INFO HandlerThread:1586 [handler.py:handle_request_defer():184] handle defer: 13 +2024-05-23 13:05:02,535 DEBUG SenderThread:1586 [sender.py:send_request():405] send_request: defer +2024-05-23 13:05:02,535 INFO SenderThread:1586 [sender.py:send_request_defer():609] handle sender defer: 13 +2024-05-23 13:05:02,535 INFO SenderThread:1586 [sender.py:transition_state():613] send defer: 14 +2024-05-23 13:05:02,535 DEBUG HandlerThread:1586 [handler.py:handle_request():158] handle_request: defer +2024-05-23 13:05:02,535 INFO HandlerThread:1586 [handler.py:handle_request_defer():184] handle defer: 14 +2024-05-23 13:05:02,536 DEBUG SenderThread:1586 [sender.py:send():378] send: final +2024-05-23 13:05:02,536 DEBUG SenderThread:1586 [sender.py:send():378] send: footer +2024-05-23 13:05:02,536 DEBUG SenderThread:1586 [sender.py:send_request():405] send_request: defer +2024-05-23 13:05:02,536 INFO SenderThread:1586 [sender.py:send_request_defer():609] handle sender defer: 14 +2024-05-23 13:05:02,536 DEBUG HandlerThread:1586 [handler.py:handle_request():158] handle_request: poll_exit +2024-05-23 13:05:02,536 DEBUG SenderThread:1586 [sender.py:send_request():405] send_request: poll_exit +2024-05-23 13:05:02,537 DEBUG HandlerThread:1586 [handler.py:handle_request():158] handle_request: poll_exit +2024-05-23 13:05:02,537 DEBUG HandlerThread:1586 [handler.py:handle_request():158] handle_request: server_info +2024-05-23 13:05:02,537 DEBUG HandlerThread:1586 [handler.py:handle_request():158] handle_request: get_summary +2024-05-23 13:05:02,537 DEBUG HandlerThread:1586 [handler.py:handle_request():158] handle_request: sampled_history +2024-05-23 13:05:02,537 DEBUG HandlerThread:1586 [handler.py:handle_request():158] handle_request: internal_messages +2024-05-23 13:05:02,537 DEBUG SenderThread:1586 [sender.py:send_request():405] send_request: poll_exit +2024-05-23 13:05:02,537 DEBUG SenderThread:1586 [sender.py:send_request():405] send_request: server_info +2024-05-23 13:05:02,593 INFO MainThread:1586 [wandb_run.py:_footer_history_summary_info():3994] rendering history +2024-05-23 13:05:02,593 INFO MainThread:1586 [wandb_run.py:_footer_history_summary_info():4026] rendering summary +2024-05-23 13:05:02,593 INFO MainThread:1586 [wandb_run.py:_footer_sync_info():3953] logging synced files +2024-05-23 13:05:02,596 DEBUG HandlerThread:1586 [handler.py:handle_request():158] handle_request: shutdown +2024-05-23 13:05:02,596 INFO HandlerThread:1586 [handler.py:finish():882] shutting down handler +2024-05-23 13:05:03,537 INFO WriterThread:1586 [datastore.py:close():296] close: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240523_130448-vm5e7ag8/run-vm5e7ag8.wandb +2024-05-23 13:05:03,593 INFO SenderThread:1586 [sender.py:finish():1545] shutting down sender +2024-05-23 13:05:03,593 INFO SenderThread:1586 [file_pusher.py:finish():169] shutting down file pusher +2024-05-23 13:05:03,593 INFO SenderThread:1586 [file_pusher.py:join():175] waiting for file pusher diff --git a/lm-evaluation-harness/wandb/run-20240523_130448-vm5e7ag8/logs/debug.log b/lm-evaluation-harness/wandb/run-20240523_130448-vm5e7ag8/logs/debug.log new file mode 100644 index 0000000000000000000000000000000000000000..bf392cfc7aca9e3063f8dc65248394e0b933b3cc --- /dev/null +++ b/lm-evaluation-harness/wandb/run-20240523_130448-vm5e7ag8/logs/debug.log @@ -0,0 +1,29 @@ +2024-05-23 13:04:48,276 INFO MainThread:1431 [wandb_setup.py:_flush():76] Current SDK version is 0.17.0 +2024-05-23 13:04:48,276 INFO MainThread:1431 [wandb_setup.py:_flush():76] Configure stats pid to 1431 +2024-05-23 13:04:48,276 INFO MainThread:1431 [wandb_setup.py:_flush():76] Loading settings from /root/.config/wandb/settings +2024-05-23 13:04:48,276 INFO MainThread:1431 [wandb_setup.py:_flush():76] Loading settings from /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/settings +2024-05-23 13:04:48,276 INFO MainThread:1431 [wandb_setup.py:_flush():76] Loading settings from environment variables: {} +2024-05-23 13:04:48,276 INFO MainThread:1431 [wandb_setup.py:_flush():76] Applying setup settings: {'_disable_service': False} +2024-05-23 13:04:48,276 WARNING MainThread:1431 [wandb_setup.py:_flush():76] Could not find program at -m lm_eval.__main__ +2024-05-23 13:04:48,276 INFO MainThread:1431 [wandb_setup.py:_flush():76] Inferring run settings from compute environment: {'program_relpath': None, 'program': '-m lm_eval.__main__'} +2024-05-23 13:04:48,276 INFO MainThread:1431 [wandb_setup.py:_flush():76] Applying login settings: {} +2024-05-23 13:04:48,276 INFO MainThread:1431 [wandb_init.py:_log_setup():520] Logging user logs to /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240523_130448-vm5e7ag8/logs/debug.log +2024-05-23 13:04:48,276 INFO MainThread:1431 [wandb_init.py:_log_setup():521] Logging internal logs to /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240523_130448-vm5e7ag8/logs/debug-internal.log +2024-05-23 13:04:48,276 INFO MainThread:1431 [wandb_init.py:init():560] calling init triggers +2024-05-23 13:04:48,276 INFO MainThread:1431 [wandb_init.py:init():567] wandb.init called with sweep_config: {} +config: {} +2024-05-23 13:04:48,276 INFO MainThread:1431 [wandb_init.py:init():610] starting backend +2024-05-23 13:04:48,276 INFO MainThread:1431 [wandb_init.py:init():614] setting up manager +2024-05-23 13:04:48,279 INFO MainThread:1431 [backend.py:_multiprocessing_setup():105] multiprocessing start_methods=fork,spawn,forkserver, using: spawn +2024-05-23 13:04:48,280 INFO MainThread:1431 [wandb_init.py:init():622] backend started and connected +2024-05-23 13:04:48,284 INFO MainThread:1431 [wandb_init.py:init():711] updated telemetry +2024-05-23 13:04:48,292 INFO MainThread:1431 [wandb_init.py:init():744] communicating run to backend with 90.0 second timeout +2024-05-23 13:04:48,542 INFO MainThread:1431 [wandb_run.py:_on_init():2396] communicating current version +2024-05-23 13:04:48,698 INFO MainThread:1431 [wandb_run.py:_on_init():2405] got version response +2024-05-23 13:04:48,698 INFO MainThread:1431 [wandb_init.py:init():795] starting run threads in backend +2024-05-23 13:04:48,977 INFO MainThread:1431 [wandb_run.py:_console_start():2374] atexit reg +2024-05-23 13:04:48,977 INFO MainThread:1431 [wandb_run.py:_redirect():2229] redirect: wrap_raw +2024-05-23 13:04:48,977 INFO MainThread:1431 [wandb_run.py:_redirect():2294] Wrapping output streams. +2024-05-23 13:04:48,977 INFO MainThread:1431 [wandb_run.py:_redirect():2319] Redirects installed. +2024-05-23 13:04:48,980 INFO MainThread:1431 [wandb_init.py:init():838] run started, returning control to user process +2024-05-23 13:05:03,597 WARNING MsgRouterThr:1431 [router.py:message_loop():77] message_loop has been closed diff --git a/lm-evaluation-harness/wandb/run-20240523_130448-vm5e7ag8/run-vm5e7ag8.wandb b/lm-evaluation-harness/wandb/run-20240523_130448-vm5e7ag8/run-vm5e7ag8.wandb new file mode 100644 index 0000000000000000000000000000000000000000..03db14f574d5b1d87d3ba86cedb26fcff49f1d06 Binary files /dev/null and b/lm-evaluation-harness/wandb/run-20240523_130448-vm5e7ag8/run-vm5e7ag8.wandb differ diff --git a/lm-evaluation-harness/wandb/run-20240530_125856-v5b29ywz/run-v5b29ywz.wandb b/lm-evaluation-harness/wandb/run-20240530_125856-v5b29ywz/run-v5b29ywz.wandb new file mode 100644 index 0000000000000000000000000000000000000000..49a411b9cb295a2bf0feee6f9710b8bedfd9b82e Binary files /dev/null and b/lm-evaluation-harness/wandb/run-20240530_125856-v5b29ywz/run-v5b29ywz.wandb differ diff --git a/venv/lib/python3.10/site-packages/transformers/models/convnextv2/__init__.py b/venv/lib/python3.10/site-packages/transformers/models/convnextv2/__init__.py new file mode 100644 index 0000000000000000000000000000000000000000..d2a484b9b8285083ba958772743207d64a8403bc --- /dev/null +++ b/venv/lib/python3.10/site-packages/transformers/models/convnextv2/__init__.py @@ -0,0 +1,97 @@ +# flake8: noqa +# There's no way to ignore "F401 '...' imported but unused" warnings in this +# module, but to preserve other warnings. So, don't check this module at all. + +# Copyright 2023 The HuggingFace Team. All rights reserved. +# +# Licensed under the Apache License, Version 2.0 (the "License"); +# you may not use this file except in compliance with the License. +# You may obtain a copy of the License at +# +# http://www.apache.org/licenses/LICENSE-2.0 +# +# Unless required by applicable law or agreed to in writing, software +# distributed under the License is distributed on an "AS IS" BASIS, +# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. +# See the License for the specific language governing permissions and +# limitations under the License. +from typing import TYPE_CHECKING + +# rely on isort to merge the imports +from ...utils import ( + OptionalDependencyNotAvailable, + _LazyModule, + is_torch_available, + is_tf_available, +) + + +_import_structure = { + "configuration_convnextv2": [ + "CONVNEXTV2_PRETRAINED_CONFIG_ARCHIVE_MAP", + "ConvNextV2Config", + ] +} + +try: + if not is_torch_available(): + raise OptionalDependencyNotAvailable() +except OptionalDependencyNotAvailable: + pass +else: + _import_structure["modeling_convnextv2"] = [ + "CONVNEXTV2_PRETRAINED_MODEL_ARCHIVE_LIST", + "ConvNextV2ForImageClassification", + "ConvNextV2Model", + "ConvNextV2PreTrainedModel", + "ConvNextV2Backbone", + ] + +try: + if not is_tf_available(): + raise OptionalDependencyNotAvailable() +except OptionalDependencyNotAvailable: + pass +else: + _import_structure["modeling_tf_convnextv2"] = [ + "TFConvNextV2ForImageClassification", + "TFConvNextV2Model", + "TFConvNextV2PreTrainedModel", + ] + +if TYPE_CHECKING: + from .configuration_convnextv2 import ( + CONVNEXTV2_PRETRAINED_CONFIG_ARCHIVE_MAP, + ConvNextV2Config, + ) + + try: + if not is_torch_available(): + raise OptionalDependencyNotAvailable() + except OptionalDependencyNotAvailable: + pass + else: + from .modeling_convnextv2 import ( + CONVNEXTV2_PRETRAINED_MODEL_ARCHIVE_LIST, + ConvNextV2Backbone, + ConvNextV2ForImageClassification, + ConvNextV2Model, + ConvNextV2PreTrainedModel, + ) + + try: + if not is_tf_available(): + raise OptionalDependencyNotAvailable() + except OptionalDependencyNotAvailable: + pass + else: + from .modeling_tf_convnextv2 import ( + TFConvNextV2ForImageClassification, + TFConvNextV2Model, + TFConvNextV2PreTrainedModel, + ) + +else: + import sys + + sys.modules[__name__] = _LazyModule(__name__, globals()["__file__"], _import_structure) diff --git a/venv/lib/python3.10/site-packages/transformers/models/convnextv2/configuration_convnextv2.py b/venv/lib/python3.10/site-packages/transformers/models/convnextv2/configuration_convnextv2.py new file mode 100644 index 0000000000000000000000000000000000000000..ccee03eef6a492332c09e523994d8bcedbcfca5f --- /dev/null +++ b/venv/lib/python3.10/site-packages/transformers/models/convnextv2/configuration_convnextv2.py @@ -0,0 +1,117 @@ +# coding=utf-8 +# Copyright 2023 Meta Platforms, Inc. and The HuggingFace Inc. team. All rights reserved. +# +# Licensed under the Apache License, Version 2.0 (the "License"); +# you may not use this file except in compliance with the License. +# You may obtain a copy of the License at +# +# http://www.apache.org/licenses/LICENSE-2.0 +# +# Unless required by applicable law or agreed to in writing, software +# distributed under the License is distributed on an "AS IS" BASIS, +# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. +# See the License for the specific language governing permissions and +# limitations under the License. +""" ConvNeXTV2 model configuration""" + + +from ...configuration_utils import PretrainedConfig +from ...utils import logging +from ...utils.backbone_utils import BackboneConfigMixin, get_aligned_output_features_output_indices + + +logger = logging.get_logger(__name__) + + +from ..deprecated._archive_maps import CONVNEXTV2_PRETRAINED_CONFIG_ARCHIVE_MAP # noqa: F401, E402 + + +class ConvNextV2Config(BackboneConfigMixin, PretrainedConfig): + r""" + This is the configuration class to store the configuration of a [`ConvNextV2Model`]. It is used to instantiate an + ConvNeXTV2 model according to the specified arguments, defining the model architecture. Instantiating a + configuration with the defaults will yield a similar configuration to that of the ConvNeXTV2 + [facebook/convnextv2-tiny-1k-224](https://huggingface.co/facebook/convnextv2-tiny-1k-224) architecture. + + Configuration objects inherit from [`PretrainedConfig`] and can be used to control the model outputs. Read the + documentation from [`PretrainedConfig`] for more information. + + Args: + num_channels (`int`, *optional*, defaults to 3): + The number of input channels. + patch_size (`int`, optional, defaults to 4): + Patch size to use in the patch embedding layer. + num_stages (`int`, optional, defaults to 4): + The number of stages in the model. + hidden_sizes (`List[int]`, *optional*, defaults to `[96, 192, 384, 768]`): + Dimensionality (hidden size) at each stage. + depths (`List[int]`, *optional*, defaults to `[3, 3, 9, 3]`): + Depth (number of blocks) for each stage. + hidden_act (`str` or `function`, *optional*, defaults to `"gelu"`): + The non-linear activation function (function or string) in each block. If string, `"gelu"`, `"relu"`, + `"selu"` and `"gelu_new"` are supported. + initializer_range (`float`, *optional*, defaults to 0.02): + The standard deviation of the truncated_normal_initializer for initializing all weight matrices. + layer_norm_eps (`float`, *optional*, defaults to 1e-12): + The epsilon used by the layer normalization layers. + drop_path_rate (`float`, *optional*, defaults to 0.0): + The drop rate for stochastic depth. + out_features (`List[str]`, *optional*): + If used as backbone, list of features to output. Can be any of `"stem"`, `"stage1"`, `"stage2"`, etc. + (depending on how many stages the model has). If unset and `out_indices` is set, will default to the + corresponding stages. If unset and `out_indices` is unset, will default to the last stage. Must be in the + same order as defined in the `stage_names` attribute. + out_indices (`List[int]`, *optional*): + If used as backbone, list of indices of features to output. Can be any of 0, 1, 2, etc. (depending on how + many stages the model has). If unset and `out_features` is set, will default to the corresponding stages. + If unset and `out_features` is unset, will default to the last stage. Must be in the + same order as defined in the `stage_names` attribute. + + Example: + ```python + >>> from transformers import ConvNeXTV2Config, ConvNextV2Model + + >>> # Initializing a ConvNeXTV2 convnextv2-tiny-1k-224 style configuration + >>> configuration = ConvNeXTV2Config() + + >>> # Initializing a model (with random weights) from the convnextv2-tiny-1k-224 style configuration + >>> model = ConvNextV2Model(configuration) + + >>> # Accessing the model configuration + >>> configuration = model.config + ```""" + + model_type = "convnextv2" + + def __init__( + self, + num_channels=3, + patch_size=4, + num_stages=4, + hidden_sizes=None, + depths=None, + hidden_act="gelu", + initializer_range=0.02, + layer_norm_eps=1e-12, + drop_path_rate=0.0, + image_size=224, + out_features=None, + out_indices=None, + **kwargs, + ): + super().__init__(**kwargs) + + self.num_channels = num_channels + self.patch_size = patch_size + self.num_stages = num_stages + self.hidden_sizes = [96, 192, 384, 768] if hidden_sizes is None else hidden_sizes + self.depths = [3, 3, 9, 3] if depths is None else depths + self.hidden_act = hidden_act + self.initializer_range = initializer_range + self.layer_norm_eps = layer_norm_eps + self.drop_path_rate = drop_path_rate + self.image_size = image_size + self.stage_names = ["stem"] + [f"stage{idx}" for idx in range(1, len(self.depths) + 1)] + self._out_features, self._out_indices = get_aligned_output_features_output_indices( + out_features=out_features, out_indices=out_indices, stage_names=self.stage_names + ) diff --git a/venv/lib/python3.10/site-packages/transformers/models/convnextv2/convert_convnextv2_to_pytorch.py b/venv/lib/python3.10/site-packages/transformers/models/convnextv2/convert_convnextv2_to_pytorch.py new file mode 100644 index 0000000000000000000000000000000000000000..8094ecf0d6157a1bb2343817f7e9303f622d9102 --- /dev/null +++ b/venv/lib/python3.10/site-packages/transformers/models/convnextv2/convert_convnextv2_to_pytorch.py @@ -0,0 +1,286 @@ +# coding=utf-8 +# Copyright 2023 The HuggingFace Inc. team. +# +# Licensed under the Apache License, Version 2.0 (the "License"); +# you may not use this file except in compliance with the License. +# You may obtain a copy of the License at +# +# http://www.apache.org/licenses/LICENSE-2.0 +# +# Unless required by applicable law or agreed to in writing, software +# distributed under the License is distributed on an "AS IS" BASIS, +# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. +# See the License for the specific language governing permissions and +# limitations under the License. +"""Convert ConvNeXTV2 checkpoints from the original repository. + +URL: https://github.com/facebookresearch/ConvNeXt""" + +import argparse +import json +import os + +import requests +import torch +from huggingface_hub import hf_hub_download +from PIL import Image + +from transformers import ConvNextImageProcessor, ConvNextV2Config, ConvNextV2ForImageClassification +from transformers.image_utils import PILImageResampling +from transformers.utils import logging + + +logging.set_verbosity_info() +logger = logging.get_logger(__name__) + + +def get_convnextv2_config(checkpoint_url): + config = ConvNextV2Config() + + if "atto" in checkpoint_url: + depths = [2, 2, 6, 2] + hidden_sizes = [40, 80, 160, 320] + if "femto" in checkpoint_url: + depths = [2, 2, 6, 2] + hidden_sizes = [48, 96, 192, 384] + if "pico" in checkpoint_url: + depths = [2, 2, 6, 2] + hidden_sizes = [64, 128, 256, 512] + if "nano" in checkpoint_url: + depths = [2, 2, 8, 2] + hidden_sizes = [80, 160, 320, 640] + if "tiny" in checkpoint_url: + depths = [3, 3, 9, 3] + hidden_sizes = [96, 192, 384, 768] + if "base" in checkpoint_url: + depths = [3, 3, 27, 3] + hidden_sizes = [128, 256, 512, 1024] + if "large" in checkpoint_url: + depths = [3, 3, 27, 3] + hidden_sizes = [192, 384, 768, 1536] + if "huge" in checkpoint_url: + depths = [3, 3, 27, 3] + hidden_sizes = [352, 704, 1408, 2816] + + num_labels = 1000 + filename = "imagenet-1k-id2label.json" + expected_shape = (1, 1000) + + repo_id = "huggingface/label-files" + config.num_labels = num_labels + id2label = json.load(open(hf_hub_download(repo_id, filename, repo_type="dataset"), "r")) + id2label = {int(k): v for k, v in id2label.items()} + + config.id2label = id2label + config.label2id = {v: k for k, v in id2label.items()} + config.hidden_sizes = hidden_sizes + config.depths = depths + + return config, expected_shape + + +def rename_key(name): + if "downsample_layers.0.0" in name: + name = name.replace("downsample_layers.0.0", "embeddings.patch_embeddings") + if "downsample_layers.0.1" in name: + name = name.replace("downsample_layers.0.1", "embeddings.norm") # we rename to layernorm later on + if "downsample_layers.1.0" in name: + name = name.replace("downsample_layers.1.0", "stages.1.downsampling_layer.0") + if "downsample_layers.1.1" in name: + name = name.replace("downsample_layers.1.1", "stages.1.downsampling_layer.1") + if "downsample_layers.2.0" in name: + name = name.replace("downsample_layers.2.0", "stages.2.downsampling_layer.0") + if "downsample_layers.2.1" in name: + name = name.replace("downsample_layers.2.1", "stages.2.downsampling_layer.1") + if "downsample_layers.3.0" in name: + name = name.replace("downsample_layers.3.0", "stages.3.downsampling_layer.0") + if "downsample_layers.3.1" in name: + name = name.replace("downsample_layers.3.1", "stages.3.downsampling_layer.1") + if "stages" in name and "downsampling_layer" not in name: + # stages.0.0. for instance should be renamed to stages.0.layers.0. + name = name[: len("stages.0")] + ".layers" + name[len("stages.0") :] + if "gamma" in name: + name = name.replace("gamma", "weight") + if "beta" in name: + name = name.replace("beta", "bias") + if "stages" in name: + name = name.replace("stages", "encoder.stages") + if "norm" in name: + name = name.replace("norm", "layernorm") + if "head" in name: + name = name.replace("head", "classifier") + + return name + + +# We will verify our results on an image of cute cats +def prepare_img(): + url = "http://images.cocodataset.org/val2017/000000039769.jpg" + im = Image.open(requests.get(url, stream=True).raw) + return im + + +def convert_preprocessor(checkpoint_url): + if "224" in checkpoint_url: + size = 224 + crop_pct = 224 / 256 + elif "384" in checkpoint_url: + size = 384 + crop_pct = None + else: + size = 512 + crop_pct = None + + return ConvNextImageProcessor( + size=size, + crop_pct=crop_pct, + image_mean=[0.485, 0.456, 0.406], + image_std=[0.229, 0.224, 0.225], + resample=PILImageResampling.BICUBIC, + ) + + +@torch.no_grad() +def convert_convnextv2_checkpoint(checkpoint_url, pytorch_dump_folder_path, save_model, push_to_hub): + """ + Copy/paste/tweak model's weights to our ConvNeXTV2 structure. + """ + print("Downloading original model from checkpoint...") + # define ConvNeXTV2 configuration based on URL + config, expected_shape = get_convnextv2_config(checkpoint_url) + # load original state_dict from URL + state_dict = torch.hub.load_state_dict_from_url(checkpoint_url)["model"] + + print("Converting model parameters...") + # rename keys + for key in state_dict.copy().keys(): + val = state_dict.pop(key) + state_dict[rename_key(key)] = val + # add prefix to all keys expect classifier head + for key in state_dict.copy().keys(): + val = state_dict.pop(key) + if not key.startswith("classifier"): + key = "convnextv2." + key + state_dict[key] = val + + # load HuggingFace model + model = ConvNextV2ForImageClassification(config) + model.load_state_dict(state_dict) + model.eval() + + # Check outputs on an image, prepared by ConvNextImageProcessor + preprocessor = convert_preprocessor(checkpoint_url) + inputs = preprocessor(images=prepare_img(), return_tensors="pt") + logits = model(**inputs).logits + + # note: the logits below were obtained without center cropping + if checkpoint_url == "https://dl.fbaipublicfiles.com/convnext/convnextv2/im1k/convnextv2_atto_1k_224_ema.pt": + expected_logits = torch.tensor([-0.3930, 0.1747, -0.5246, 0.4177, 0.4295]) + elif checkpoint_url == "https://dl.fbaipublicfiles.com/convnext/convnextv2/im1k/convnextv2_femto_1k_224_ema.pt": + expected_logits = torch.tensor([-0.1727, -0.5341, -0.7818, -0.4745, -0.6566]) + elif checkpoint_url == "https://dl.fbaipublicfiles.com/convnext/convnextv2/im1k/convnextv2_pico_1k_224_ema.pt": + expected_logits = torch.tensor([-0.0333, 0.1563, -0.9137, 0.1054, 0.0381]) + elif checkpoint_url == "https://dl.fbaipublicfiles.com/convnext/convnextv2/im1k/convnextv2_nano_1k_224_ema.pt": + expected_logits = torch.tensor([-0.1744, -0.1555, -0.0713, 0.0950, -0.1431]) + elif checkpoint_url == "https://dl.fbaipublicfiles.com/convnext/convnextv2/im1k/convnextv2_tiny_1k_224_ema.pt": + expected_logits = torch.tensor([0.9996, 0.1966, -0.4386, -0.3472, 0.6661]) + elif checkpoint_url == "https://dl.fbaipublicfiles.com/convnext/convnextv2/im1k/convnextv2_base_1k_224_ema.pt": + expected_logits = torch.tensor([-0.2553, -0.6708, -0.1359, 0.2518, -0.2488]) + elif checkpoint_url == "https://dl.fbaipublicfiles.com/convnext/convnextv2/im1k/convnextv2_large_1k_224_ema.pt": + expected_logits = torch.tensor([-0.0673, -0.5627, -0.3753, -0.2722, 0.0178]) + elif checkpoint_url == "https://dl.fbaipublicfiles.com/convnext/convnextv2/im1k/convnextv2_huge_1k_224_ema.pt": + expected_logits = torch.tensor([-0.6377, -0.7458, -0.2150, 0.1184, -0.0597]) + elif checkpoint_url == "https://dl.fbaipublicfiles.com/convnext/convnextv2/im22k/convnextv2_nano_22k_224_ema.pt": + expected_logits = torch.tensor([1.0799, 0.2322, -0.8860, 1.0219, 0.6231]) + elif checkpoint_url == "https://dl.fbaipublicfiles.com/convnext/convnextv2/im22k/convnextv2_nano_22k_384_ema.pt": + expected_logits = torch.tensor([0.3766, 0.4917, -1.1426, 0.9942, 0.6024]) + elif checkpoint_url == "https://dl.fbaipublicfiles.com/convnext/convnextv2/im22k/convnextv2_tiny_22k_224_ema.pt": + expected_logits = torch.tensor([0.4220, -0.6919, -0.4317, -0.2881, -0.6609]) + elif checkpoint_url == "https://dl.fbaipublicfiles.com/convnext/convnextv2/im22k/convnextv2_tiny_22k_384_ema.pt": + expected_logits = torch.tensor([0.1082, -0.8286, -0.5095, 0.4681, -0.8085]) + elif checkpoint_url == "https://dl.fbaipublicfiles.com/convnext/convnextv2/im22k/convnextv2_base_22k_224_ema.pt": + expected_logits = torch.tensor([-0.2419, -0.6221, 0.2176, -0.0980, -0.7527]) + elif checkpoint_url == "https://dl.fbaipublicfiles.com/convnext/convnextv2/im22k/convnextv2_base_22k_384_ema.pt": + expected_logits = torch.tensor([0.0391, -0.4371, 0.3786, 0.1251, -0.2784]) + elif checkpoint_url == "https://dl.fbaipublicfiles.com/convnext/convnextv2/im22k/convnextv2_large_22k_224_ema.pt": + expected_logits = torch.tensor([-0.0504, 0.5636, -0.1729, -0.6507, -0.3949]) + elif checkpoint_url == "https://dl.fbaipublicfiles.com/convnext/convnextv2/im22k/convnextv2_large_22k_384_ema.pt": + expected_logits = torch.tensor([0.3560, 0.9486, 0.3149, -0.2667, -0.5138]) + elif checkpoint_url == "https://dl.fbaipublicfiles.com/convnext/convnextv2/im22k/convnextv2_huge_22k_384_ema.pt": + expected_logits = torch.tensor([-0.2469, -0.4550, -0.5853, -0.0810, 0.0309]) + elif checkpoint_url == "https://dl.fbaipublicfiles.com/convnext/convnextv2/im22k/convnextv2_huge_22k_512_ema.pt": + expected_logits = torch.tensor([-0.3090, 0.0802, -0.0682, -0.1979, -0.2826]) + else: + raise ValueError(f"Unknown URL: {checkpoint_url}") + + assert torch.allclose(logits[0, :5], expected_logits, atol=1e-3) + assert logits.shape == expected_shape + print("Model outputs match the original results!") + + if save_model: + print("Saving model to local...") + # Create folder to save model + if not os.path.isdir(pytorch_dump_folder_path): + os.mkdir(pytorch_dump_folder_path) + + model.save_pretrained(pytorch_dump_folder_path) + preprocessor.save_pretrained(pytorch_dump_folder_path) + + model_name = "convnextv2" + if "atto" in checkpoint_url: + model_name += "-atto" + if "femto" in checkpoint_url: + model_name += "-femto" + if "pico" in checkpoint_url: + model_name += "-pico" + if "nano" in checkpoint_url: + model_name += "-nano" + elif "tiny" in checkpoint_url: + model_name += "-tiny" + elif "base" in checkpoint_url: + model_name += "-base" + elif "large" in checkpoint_url: + model_name += "-large" + elif "huge" in checkpoint_url: + model_name += "-huge" + if "22k" in checkpoint_url and "1k" not in checkpoint_url: + model_name += "-22k" + elif "22k" in checkpoint_url and "1k" in checkpoint_url: + model_name += "-22k-1k" + elif "1k" in checkpoint_url: + model_name += "-1k" + if "224" in checkpoint_url: + model_name += "-224" + elif "384" in checkpoint_url: + model_name += "-384" + elif "512" in checkpoint_url: + model_name += "-512" + + if push_to_hub: + print(f"Pushing {model_name} to the hub...") + model.push_to_hub(model_name) + preprocessor.push_to_hub(model_name) + + +if __name__ == "__main__": + parser = argparse.ArgumentParser() + # Required parameters + parser.add_argument( + "--checkpoint_url", + default="https://dl.fbaipublicfiles.com/convnext/convnextv2/im1k/convnextv2_atto_1k_224_ema.pt", + type=str, + help="URL of the original ConvNeXTV2 checkpoint you'd like to convert.", + ) + parser.add_argument( + "--pytorch_dump_folder_path", + default="model", + type=str, + help="Path to the output PyTorch model directory.", + ) + parser.add_argument("--save_model", action="store_true", help="Save model to local") + parser.add_argument("--push_to_hub", action="store_true", help="Push model and image preprocessor to the hub") + + args = parser.parse_args() + convert_convnextv2_checkpoint( + args.checkpoint_url, args.pytorch_dump_folder_path, args.save_model, args.push_to_hub + ) diff --git a/venv/lib/python3.10/site-packages/transformers/models/convnextv2/modeling_convnextv2.py b/venv/lib/python3.10/site-packages/transformers/models/convnextv2/modeling_convnextv2.py new file mode 100644 index 0000000000000000000000000000000000000000..7439f212971ec11c21215cbecb665f21405145f0 --- /dev/null +++ b/venv/lib/python3.10/site-packages/transformers/models/convnextv2/modeling_convnextv2.py @@ -0,0 +1,574 @@ +# coding=utf-8 +# Copyright 2023 Meta Platforms, Inc. and The HuggingFace Inc. team. All rights reserved. +# +# Licensed under the Apache License, Version 2.0 (the "License"); +# you may not use this file except in compliance with the License. +# You may obtain a copy of the License at +# +# http://www.apache.org/licenses/LICENSE-2.0 +# +# Unless required by applicable law or agreed to in writing, software +# distributed under the License is distributed on an "AS IS" BASIS, +# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. +# See the License for the specific language governing permissions and +# limitations under the License. +""" PyTorch ConvNextV2 model.""" + + +from typing import Optional, Tuple, Union + +import torch +import torch.utils.checkpoint +from torch import nn +from torch.nn import BCEWithLogitsLoss, CrossEntropyLoss, MSELoss + +from ...activations import ACT2FN +from ...modeling_outputs import ( + BackboneOutput, + BaseModelOutputWithNoAttention, + BaseModelOutputWithPoolingAndNoAttention, + ImageClassifierOutputWithNoAttention, +) +from ...modeling_utils import PreTrainedModel +from ...utils import ( + add_code_sample_docstrings, + add_start_docstrings, + add_start_docstrings_to_model_forward, + logging, + replace_return_docstrings, +) +from ...utils.backbone_utils import BackboneMixin +from .configuration_convnextv2 import ConvNextV2Config + + +logger = logging.get_logger(__name__) + +# General docstring +_CONFIG_FOR_DOC = "ConvNextV2Config" + +# Base docstring +_CHECKPOINT_FOR_DOC = "facebook/convnextv2-tiny-1k-224" +_EXPECTED_OUTPUT_SHAPE = [1, 768, 7, 7] + +# Image classification docstring +_IMAGE_CLASS_CHECKPOINT = "facebook/convnextv2-tiny-1k-224" +_IMAGE_CLASS_EXPECTED_OUTPUT = "tabby, tabby cat" + + +from ..deprecated._archive_maps import CONVNEXTV2_PRETRAINED_MODEL_ARCHIVE_LIST # noqa: F401, E402 + + +# Copied from transformers.models.beit.modeling_beit.drop_path +def drop_path(input: torch.Tensor, drop_prob: float = 0.0, training: bool = False) -> torch.Tensor: + """ + Drop paths (Stochastic Depth) per sample (when applied in main path of residual blocks). + + Comment by Ross Wightman: This is the same as the DropConnect impl I created for EfficientNet, etc networks, + however, the original name is misleading as 'Drop Connect' is a different form of dropout in a separate paper... + See discussion: https://github.com/tensorflow/tpu/issues/494#issuecomment-532968956 ... I've opted for changing the + layer and argument names to 'drop path' rather than mix DropConnect as a layer name and use 'survival rate' as the + argument. + """ + if drop_prob == 0.0 or not training: + return input + keep_prob = 1 - drop_prob + shape = (input.shape[0],) + (1,) * (input.ndim - 1) # work with diff dim tensors, not just 2D ConvNets + random_tensor = keep_prob + torch.rand(shape, dtype=input.dtype, device=input.device) + random_tensor.floor_() # binarize + output = input.div(keep_prob) * random_tensor + return output + + +# Copied from transformers.models.beit.modeling_beit.BeitDropPath with Beit->ConvNextV2 +class ConvNextV2DropPath(nn.Module): + """Drop paths (Stochastic Depth) per sample (when applied in main path of residual blocks).""" + + def __init__(self, drop_prob: Optional[float] = None) -> None: + super().__init__() + self.drop_prob = drop_prob + + def forward(self, hidden_states: torch.Tensor) -> torch.Tensor: + return drop_path(hidden_states, self.drop_prob, self.training) + + def extra_repr(self) -> str: + return "p={}".format(self.drop_prob) + + +class ConvNextV2GRN(nn.Module): + """GRN (Global Response Normalization) layer""" + + def __init__(self, dim: int): + super().__init__() + self.weight = nn.Parameter(torch.zeros(1, 1, 1, dim)) + self.bias = nn.Parameter(torch.zeros(1, 1, 1, dim)) + + def forward(self, hidden_states: torch.FloatTensor) -> torch.FloatTensor: + # Compute and normalize global spatial feature maps + global_features = torch.norm(hidden_states, p=2, dim=(1, 2), keepdim=True) + norm_features = global_features / (global_features.mean(dim=-1, keepdim=True) + 1e-6) + hidden_states = self.weight * (hidden_states * norm_features) + self.bias + hidden_states + + return hidden_states + + +# Copied from transformers.models.convnext.modeling_convnext.ConvNextLayerNorm with ConvNext->ConvNextV2 +class ConvNextV2LayerNorm(nn.Module): + r"""LayerNorm that supports two data formats: channels_last (default) or channels_first. + The ordering of the dimensions in the inputs. channels_last corresponds to inputs with shape (batch_size, height, + width, channels) while channels_first corresponds to inputs with shape (batch_size, channels, height, width). + """ + + def __init__(self, normalized_shape, eps=1e-6, data_format="channels_last"): + super().__init__() + self.weight = nn.Parameter(torch.ones(normalized_shape)) + self.bias = nn.Parameter(torch.zeros(normalized_shape)) + self.eps = eps + self.data_format = data_format + if self.data_format not in ["channels_last", "channels_first"]: + raise NotImplementedError(f"Unsupported data format: {self.data_format}") + self.normalized_shape = (normalized_shape,) + + def forward(self, x: torch.Tensor) -> torch.Tensor: + if self.data_format == "channels_last": + x = torch.nn.functional.layer_norm(x, self.normalized_shape, self.weight, self.bias, self.eps) + elif self.data_format == "channels_first": + input_dtype = x.dtype + x = x.float() + u = x.mean(1, keepdim=True) + s = (x - u).pow(2).mean(1, keepdim=True) + x = (x - u) / torch.sqrt(s + self.eps) + x = x.to(dtype=input_dtype) + x = self.weight[:, None, None] * x + self.bias[:, None, None] + return x + + +# Copied from transformers.models.convnext.modeling_convnext.ConvNextEmbeddings with ConvNext->ConvNextV2 +class ConvNextV2Embeddings(nn.Module): + """This class is comparable to (and inspired by) the SwinEmbeddings class + found in src/transformers/models/swin/modeling_swin.py. + """ + + def __init__(self, config): + super().__init__() + self.patch_embeddings = nn.Conv2d( + config.num_channels, config.hidden_sizes[0], kernel_size=config.patch_size, stride=config.patch_size + ) + self.layernorm = ConvNextV2LayerNorm(config.hidden_sizes[0], eps=1e-6, data_format="channels_first") + self.num_channels = config.num_channels + + def forward(self, pixel_values: torch.FloatTensor) -> torch.Tensor: + num_channels = pixel_values.shape[1] + if num_channels != self.num_channels: + raise ValueError( + "Make sure that the channel dimension of the pixel values match with the one set in the configuration." + ) + embeddings = self.patch_embeddings(pixel_values) + embeddings = self.layernorm(embeddings) + return embeddings + + +class ConvNextV2Layer(nn.Module): + """This corresponds to the `Block` class in the original implementation. + + There are two equivalent implementations: [DwConv, LayerNorm (channels_first), Conv, GELU,1x1 Conv]; all in (N, C, + H, W) (2) [DwConv, Permute to (N, H, W, C), LayerNorm (channels_last), Linear, GELU, Linear]; Permute back + + The authors used (2) as they find it slightly faster in PyTorch. + + Args: + config ([`ConvNextV2Config`]): Model configuration class. + dim (`int`): Number of input channels. + drop_path (`float`): Stochastic depth rate. Default: 0.0. + """ + + def __init__(self, config, dim, drop_path=0): + super().__init__() + # depthwise conv + self.dwconv = nn.Conv2d(dim, dim, kernel_size=7, padding=3, groups=dim) + self.layernorm = ConvNextV2LayerNorm(dim, eps=1e-6) + # pointwise/1x1 convs, implemented with linear layers + self.pwconv1 = nn.Linear(dim, 4 * dim) + self.act = ACT2FN[config.hidden_act] + self.grn = ConvNextV2GRN(4 * dim) + self.pwconv2 = nn.Linear(4 * dim, dim) + self.drop_path = ConvNextV2DropPath(drop_path) if drop_path > 0.0 else nn.Identity() + + def forward(self, hidden_states: torch.FloatTensor) -> torch.Tensor: + input = hidden_states + x = self.dwconv(hidden_states) + # (batch_size, num_channels, height, width) -> (batch_size, height, width, num_channels) + x = x.permute(0, 2, 3, 1) + x = self.layernorm(x) + x = self.pwconv1(x) + x = self.act(x) + x = self.grn(x) + x = self.pwconv2(x) + # (batch_size, height, width, num_channels) -> (batch_size, num_channels, height, width) + x = x.permute(0, 3, 1, 2) + + x = input + self.drop_path(x) + return x + + +# Copied from transformers.models.convnext.modeling_convnext.ConvNextStage with ConvNeXT->ConvNeXTV2, ConvNext->ConvNextV2 +class ConvNextV2Stage(nn.Module): + """ConvNeXTV2 stage, consisting of an optional downsampling layer + multiple residual blocks. + + Args: + config ([`ConvNextV2Config`]): Model configuration class. + in_channels (`int`): Number of input channels. + out_channels (`int`): Number of output channels. + depth (`int`): Number of residual blocks. + drop_path_rates(`List[float]`): Stochastic depth rates for each layer. + """ + + def __init__(self, config, in_channels, out_channels, kernel_size=2, stride=2, depth=2, drop_path_rates=None): + super().__init__() + + if in_channels != out_channels or stride > 1: + self.downsampling_layer = nn.Sequential( + ConvNextV2LayerNorm(in_channels, eps=1e-6, data_format="channels_first"), + nn.Conv2d(in_channels, out_channels, kernel_size=kernel_size, stride=stride), + ) + else: + self.downsampling_layer = nn.Identity() + drop_path_rates = drop_path_rates or [0.0] * depth + self.layers = nn.Sequential( + *[ConvNextV2Layer(config, dim=out_channels, drop_path=drop_path_rates[j]) for j in range(depth)] + ) + + def forward(self, hidden_states: torch.FloatTensor) -> torch.Tensor: + hidden_states = self.downsampling_layer(hidden_states) + hidden_states = self.layers(hidden_states) + return hidden_states + + +# Copied from transformers.models.convnext.modeling_convnext.ConvNextEncoder with ConvNext->ConvNextV2 +class ConvNextV2Encoder(nn.Module): + def __init__(self, config): + super().__init__() + self.stages = nn.ModuleList() + drop_path_rates = [ + x.tolist() for x in torch.linspace(0, config.drop_path_rate, sum(config.depths)).split(config.depths) + ] + prev_chs = config.hidden_sizes[0] + for i in range(config.num_stages): + out_chs = config.hidden_sizes[i] + stage = ConvNextV2Stage( + config, + in_channels=prev_chs, + out_channels=out_chs, + stride=2 if i > 0 else 1, + depth=config.depths[i], + drop_path_rates=drop_path_rates[i], + ) + self.stages.append(stage) + prev_chs = out_chs + + def forward( + self, + hidden_states: torch.FloatTensor, + output_hidden_states: Optional[bool] = False, + return_dict: Optional[bool] = True, + ) -> Union[Tuple, BaseModelOutputWithNoAttention]: + all_hidden_states = () if output_hidden_states else None + + for i, layer_module in enumerate(self.stages): + if output_hidden_states: + all_hidden_states = all_hidden_states + (hidden_states,) + + hidden_states = layer_module(hidden_states) + + if output_hidden_states: + all_hidden_states = all_hidden_states + (hidden_states,) + + if not return_dict: + return tuple(v for v in [hidden_states, all_hidden_states] if v is not None) + + return BaseModelOutputWithNoAttention( + last_hidden_state=hidden_states, + hidden_states=all_hidden_states, + ) + + +# Copied from transformers.models.convnext.modeling_convnext.ConvNextPreTrainedModel with ConvNext->ConvNextV2, convnext->convnextv2 +class ConvNextV2PreTrainedModel(PreTrainedModel): + """ + An abstract class to handle weights initialization and a simple interface for downloading and loading pretrained + models. + """ + + config_class = ConvNextV2Config + base_model_prefix = "convnextv2" + main_input_name = "pixel_values" + + def _init_weights(self, module): + """Initialize the weights""" + if isinstance(module, (nn.Linear, nn.Conv2d)): + # Slightly different from the TF version which uses truncated_normal for initialization + # cf https://github.com/pytorch/pytorch/pull/5617 + module.weight.data.normal_(mean=0.0, std=self.config.initializer_range) + if module.bias is not None: + module.bias.data.zero_() + elif isinstance(module, nn.LayerNorm): + module.bias.data.zero_() + module.weight.data.fill_(1.0) + + +CONVNEXTV2_START_DOCSTRING = r""" + This model is a PyTorch [torch.nn.Module](https://pytorch.org/docs/stable/nn.html#torch.nn.Module) subclass. Use it + as a regular PyTorch Module and refer to the PyTorch documentation for all matter related to general usage and + behavior. + + Parameters: + config ([`ConvNextV2Config`]): Model configuration class with all the parameters of the model. + Initializing with a config file does not load the weights associated with the model, only the + configuration. Check out the [`~PreTrainedModel.from_pretrained`] method to load the model weights. +""" + +CONVNEXTV2_INPUTS_DOCSTRING = r""" + Args: + pixel_values (`torch.FloatTensor` of shape `(batch_size, num_channels, height, width)`): + Pixel values. Pixel values can be obtained using [`ConvNextImageProcessor`]. See + [`ConvNextImageProcessor.__call__`] for details. + output_hidden_states (`bool`, *optional*): + Whether or not to return the hidden states of all layers. See `hidden_states` under returned tensors for + more detail. + return_dict (`bool`, *optional*): + Whether or not to return a [`~utils.ModelOutput`] instead of a plain tuple. +""" + + +@add_start_docstrings( + "The bare ConvNextV2 model outputting raw features without any specific head on top.", + CONVNEXTV2_START_DOCSTRING, +) +# Copied from transformers.models.convnext.modeling_convnext.ConvNextModel with CONVNEXT->CONVNEXTV2, ConvNext->ConvNextV2 +class ConvNextV2Model(ConvNextV2PreTrainedModel): + def __init__(self, config): + super().__init__(config) + self.config = config + + self.embeddings = ConvNextV2Embeddings(config) + self.encoder = ConvNextV2Encoder(config) + + # final layernorm layer + self.layernorm = nn.LayerNorm(config.hidden_sizes[-1], eps=config.layer_norm_eps) + + # Initialize weights and apply final processing + self.post_init() + + @add_start_docstrings_to_model_forward(CONVNEXTV2_INPUTS_DOCSTRING) + @add_code_sample_docstrings( + checkpoint=_CHECKPOINT_FOR_DOC, + output_type=BaseModelOutputWithPoolingAndNoAttention, + config_class=_CONFIG_FOR_DOC, + modality="vision", + expected_output=_EXPECTED_OUTPUT_SHAPE, + ) + def forward( + self, + pixel_values: torch.FloatTensor = None, + output_hidden_states: Optional[bool] = None, + return_dict: Optional[bool] = None, + ) -> Union[Tuple, BaseModelOutputWithPoolingAndNoAttention]: + output_hidden_states = ( + output_hidden_states if output_hidden_states is not None else self.config.output_hidden_states + ) + return_dict = return_dict if return_dict is not None else self.config.use_return_dict + + if pixel_values is None: + raise ValueError("You have to specify pixel_values") + + embedding_output = self.embeddings(pixel_values) + + encoder_outputs = self.encoder( + embedding_output, + output_hidden_states=output_hidden_states, + return_dict=return_dict, + ) + + last_hidden_state = encoder_outputs[0] + + # global average pooling, (N, C, H, W) -> (N, C) + pooled_output = self.layernorm(last_hidden_state.mean([-2, -1])) + + if not return_dict: + return (last_hidden_state, pooled_output) + encoder_outputs[1:] + + return BaseModelOutputWithPoolingAndNoAttention( + last_hidden_state=last_hidden_state, + pooler_output=pooled_output, + hidden_states=encoder_outputs.hidden_states, + ) + + +@add_start_docstrings( + """ + ConvNextV2 Model with an image classification head on top (a linear layer on top of the pooled features), e.g. for + ImageNet. + """, + CONVNEXTV2_START_DOCSTRING, +) +# Copied from transformers.models.convnext.modeling_convnext.ConvNextForImageClassification with CONVNEXT->CONVNEXTV2,ConvNext->ConvNextV2,convnext->convnextv2 +class ConvNextV2ForImageClassification(ConvNextV2PreTrainedModel): + def __init__(self, config): + super().__init__(config) + + self.num_labels = config.num_labels + self.convnextv2 = ConvNextV2Model(config) + + # Classifier head + self.classifier = ( + nn.Linear(config.hidden_sizes[-1], config.num_labels) if config.num_labels > 0 else nn.Identity() + ) + + # Initialize weights and apply final processing + self.post_init() + + @add_start_docstrings_to_model_forward(CONVNEXTV2_INPUTS_DOCSTRING) + @add_code_sample_docstrings( + checkpoint=_IMAGE_CLASS_CHECKPOINT, + output_type=ImageClassifierOutputWithNoAttention, + config_class=_CONFIG_FOR_DOC, + expected_output=_IMAGE_CLASS_EXPECTED_OUTPUT, + ) + def forward( + self, + pixel_values: torch.FloatTensor = None, + labels: Optional[torch.LongTensor] = None, + output_hidden_states: Optional[bool] = None, + return_dict: Optional[bool] = None, + ) -> Union[Tuple, ImageClassifierOutputWithNoAttention]: + r""" + labels (`torch.LongTensor` of shape `(batch_size,)`, *optional*): + Labels for computing the image classification/regression loss. Indices should be in `[0, ..., + config.num_labels - 1]`. If `config.num_labels == 1` a regression loss is computed (Mean-Square loss), If + `config.num_labels > 1` a classification loss is computed (Cross-Entropy). + """ + return_dict = return_dict if return_dict is not None else self.config.use_return_dict + + outputs = self.convnextv2(pixel_values, output_hidden_states=output_hidden_states, return_dict=return_dict) + + pooled_output = outputs.pooler_output if return_dict else outputs[1] + + logits = self.classifier(pooled_output) + + loss = None + if labels is not None: + if self.config.problem_type is None: + if self.num_labels == 1: + self.config.problem_type = "regression" + elif self.num_labels > 1 and (labels.dtype == torch.long or labels.dtype == torch.int): + self.config.problem_type = "single_label_classification" + else: + self.config.problem_type = "multi_label_classification" + + if self.config.problem_type == "regression": + loss_fct = MSELoss() + if self.num_labels == 1: + loss = loss_fct(logits.squeeze(), labels.squeeze()) + else: + loss = loss_fct(logits, labels) + elif self.config.problem_type == "single_label_classification": + loss_fct = CrossEntropyLoss() + loss = loss_fct(logits.view(-1, self.num_labels), labels.view(-1)) + elif self.config.problem_type == "multi_label_classification": + loss_fct = BCEWithLogitsLoss() + loss = loss_fct(logits, labels) + if not return_dict: + output = (logits,) + outputs[2:] + return ((loss,) + output) if loss is not None else output + + return ImageClassifierOutputWithNoAttention( + loss=loss, + logits=logits, + hidden_states=outputs.hidden_states, + ) + + +@add_start_docstrings( + """ + ConvNeXT V2 backbone, to be used with frameworks like DETR and MaskFormer. + """, + CONVNEXTV2_START_DOCSTRING, +) +# Copied from transformers.models.convnext.modeling_convnext.ConvNextBackbone with CONVNEXT->CONVNEXTV2,ConvNext->ConvNextV2,facebook/convnext-tiny-224->facebook/convnextv2-tiny-1k-224 +class ConvNextV2Backbone(ConvNextV2PreTrainedModel, BackboneMixin): + def __init__(self, config): + super().__init__(config) + super()._init_backbone(config) + + self.embeddings = ConvNextV2Embeddings(config) + self.encoder = ConvNextV2Encoder(config) + self.num_features = [config.hidden_sizes[0]] + config.hidden_sizes + + # Add layer norms to hidden states of out_features + hidden_states_norms = {} + for stage, num_channels in zip(self._out_features, self.channels): + hidden_states_norms[stage] = ConvNextV2LayerNorm(num_channels, data_format="channels_first") + self.hidden_states_norms = nn.ModuleDict(hidden_states_norms) + + # initialize weights and apply final processing + self.post_init() + + @add_start_docstrings_to_model_forward(CONVNEXTV2_INPUTS_DOCSTRING) + @replace_return_docstrings(output_type=BackboneOutput, config_class=_CONFIG_FOR_DOC) + def forward( + self, + pixel_values: torch.Tensor, + output_hidden_states: Optional[bool] = None, + return_dict: Optional[bool] = None, + ) -> BackboneOutput: + """ + Returns: + + Examples: + + ```python + >>> from transformers import AutoImageProcessor, AutoBackbone + >>> import torch + >>> from PIL import Image + >>> import requests + + >>> url = "http://images.cocodataset.org/val2017/000000039769.jpg" + >>> image = Image.open(requests.get(url, stream=True).raw) + + >>> processor = AutoImageProcessor.from_pretrained("facebook/convnextv2-tiny-1k-224") + >>> model = AutoBackbone.from_pretrained("facebook/convnextv2-tiny-1k-224") + + >>> inputs = processor(image, return_tensors="pt") + >>> outputs = model(**inputs) + ```""" + return_dict = return_dict if return_dict is not None else self.config.use_return_dict + output_hidden_states = ( + output_hidden_states if output_hidden_states is not None else self.config.output_hidden_states + ) + + embedding_output = self.embeddings(pixel_values) + + outputs = self.encoder( + embedding_output, + output_hidden_states=True, + return_dict=return_dict, + ) + + hidden_states = outputs.hidden_states if return_dict else outputs[1] + + feature_maps = () + for stage, hidden_state in zip(self.stage_names, hidden_states): + if stage in self.out_features: + hidden_state = self.hidden_states_norms[stage](hidden_state) + feature_maps += (hidden_state,) + + if not return_dict: + output = (feature_maps,) + if output_hidden_states: + output += (hidden_states,) + return output + + return BackboneOutput( + feature_maps=feature_maps, + hidden_states=hidden_states if output_hidden_states else None, + attentions=None, + ) diff --git a/venv/lib/python3.10/site-packages/transformers/models/convnextv2/modeling_tf_convnextv2.py b/venv/lib/python3.10/site-packages/transformers/models/convnextv2/modeling_tf_convnextv2.py new file mode 100644 index 0000000000000000000000000000000000000000..0debe6fd0c54d6387bcb471370585472dc9d6921 --- /dev/null +++ b/venv/lib/python3.10/site-packages/transformers/models/convnextv2/modeling_tf_convnextv2.py @@ -0,0 +1,681 @@ +# coding=utf-8 +# Copyright 2023 Meta Platforms Inc. and The HuggingFace Inc. team. All rights reserved. +# +# Licensed under the Apache License, Version 2.0 (the "License"); +# you may not use this file except in compliance with the License. +# You may obtain a copy of the License at +# +# http://www.apache.org/licenses/LICENSE-2.0 +# +# Unless required by applicable law or agreed to in writing, software +# distributed under the License is distributed on an "AS IS" BASIS, +# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. +# See the License for the specific language governing permissions and +# limitations under the License. +""" TF 2.0 ConvNextV2 model.""" + + +from __future__ import annotations + +from typing import List, Optional, Tuple, Union + +import numpy as np +import tensorflow as tf + +from ...activations_tf import get_tf_activation +from ...modeling_tf_outputs import ( + TFBaseModelOutputWithNoAttention, + TFBaseModelOutputWithPooling, + TFBaseModelOutputWithPoolingAndNoAttention, + TFImageClassifierOutputWithNoAttention, +) +from ...modeling_tf_utils import ( + TFModelInputType, + TFPreTrainedModel, + TFSequenceClassificationLoss, + get_initializer, + keras, + keras_serializable, + unpack_inputs, +) +from ...tf_utils import shape_list +from ...utils import ( + add_code_sample_docstrings, + add_start_docstrings, + add_start_docstrings_to_model_forward, + logging, +) +from .configuration_convnextv2 import ConvNextV2Config + + +logger = logging.get_logger(__name__) + +# General docstring +_CONFIG_FOR_DOC = "ConvNextV2Config" + +# Base docstring +_CHECKPOINT_FOR_DOC = "facebook/convnextv2-tiny-1k-224" +_EXPECTED_OUTPUT_SHAPE = [1, 768, 7, 7] + +# Image classification docstring +_IMAGE_CLASS_CHECKPOINT = "facebook/convnextv2-tiny-1k-224" +_IMAGE_CLASS_EXPECTED_OUTPUT = "tabby, tabby cat" + + +# Copied from transformers.models.convnext.modeling_tf_convnext.TFConvNextDropPath with ConvNext->ConvNextV2 +class TFConvNextV2DropPath(keras.layers.Layer): + """Drop paths (Stochastic Depth) per sample (when applied in main path of residual blocks). + References: + (1) github.com:rwightman/pytorch-image-models + """ + + def __init__(self, drop_path: float, **kwargs): + super().__init__(**kwargs) + self.drop_path = drop_path + + def call(self, x: tf.Tensor, training=None): + if training: + keep_prob = 1 - self.drop_path + shape = (tf.shape(x)[0],) + (1,) * (len(tf.shape(x)) - 1) + random_tensor = keep_prob + tf.random.uniform(shape, 0, 1) + random_tensor = tf.floor(random_tensor) + return (x / keep_prob) * random_tensor + return x + + +class TFConvNextV2GRN(keras.layers.Layer): + """GRN (Global Response Normalization) layer""" + + def __init__(self, config: ConvNextV2Config, dim: int, **kwargs): + super().__init__(**kwargs) + self.dim = dim + + def build(self, input_shape: tf.TensorShape = None): + # PT's `nn.Parameters` must be mapped to a TF layer weight to inherit the same name hierarchy (and vice-versa) + self.weight = self.add_weight( + name="weight", + shape=(1, 1, 1, self.dim), + initializer=keras.initializers.Zeros(), + ) + self.bias = self.add_weight( + name="bias", + shape=(1, 1, 1, self.dim), + initializer=keras.initializers.Zeros(), + ) + return super().build(input_shape) + + def call(self, hidden_states: tf.Tensor): + global_features = tf.norm(hidden_states, ord="euclidean", axis=(1, 2), keepdims=True) + norm_features = global_features / (tf.reduce_mean(global_features, axis=-1, keepdims=True) + 1e-6) + hidden_states = self.weight * (hidden_states * norm_features) + self.bias + hidden_states + return hidden_states + + +# Copied from transformers.models.convnext.modeling_tf_convnext.TFConvNextEmbeddings with ConvNext->ConvNextV2 +class TFConvNextV2Embeddings(keras.layers.Layer): + """This class is comparable to (and inspired by) the SwinEmbeddings class + found in src/transformers/models/swin/modeling_swin.py. + """ + + def __init__(self, config: ConvNextV2Config, **kwargs): + super().__init__(**kwargs) + self.patch_embeddings = keras.layers.Conv2D( + filters=config.hidden_sizes[0], + kernel_size=config.patch_size, + strides=config.patch_size, + name="patch_embeddings", + kernel_initializer=get_initializer(config.initializer_range), + bias_initializer=keras.initializers.Zeros(), + ) + self.layernorm = keras.layers.LayerNormalization(epsilon=1e-6, name="layernorm") + self.num_channels = config.num_channels + self.config = config + + def call(self, pixel_values): + if isinstance(pixel_values, dict): + pixel_values = pixel_values["pixel_values"] + + tf.debugging.assert_equal( + shape_list(pixel_values)[1], + self.num_channels, + message="Make sure that the channel dimension of the pixel values match with the one set in the configuration.", + ) + + # When running on CPU, `keras.layers.Conv2D` doesn't support `NCHW` format. + # So change the input format from `NCHW` to `NHWC`. + # shape = (batch_size, in_height, in_width, in_channels) + pixel_values = tf.transpose(pixel_values, perm=(0, 2, 3, 1)) + + embeddings = self.patch_embeddings(pixel_values) + embeddings = self.layernorm(embeddings) + return embeddings + + def build(self, input_shape=None): + if self.built: + return + self.built = True + if getattr(self, "patch_embeddings", None) is not None: + with tf.name_scope(self.patch_embeddings.name): + self.patch_embeddings.build([None, None, None, self.config.num_channels]) + if getattr(self, "layernorm", None) is not None: + with tf.name_scope(self.layernorm.name): + self.layernorm.build([None, None, None, self.config.hidden_sizes[0]]) + + +class TFConvNextV2Layer(keras.layers.Layer): + """This corresponds to the `Block` class in the original implementation. + + There are two equivalent implementations: [DwConv, LayerNorm (channels_first), Conv, GELU,1x1 Conv]; all in (N, C, + H, W) (2) [DwConv, Permute to (N, H, W, C), LayerNorm (channels_last), Linear, GELU, Linear]; Permute back + + The authors used (2) as they find it slightly faster in PyTorch. Since we already permuted the inputs to follow + NHWC ordering, we can just apply the operations straight-away without the permutation. + + Args: + config (`ConvNextV2Config`): + Model configuration class. + dim (`int`): + Number of input channels. + drop_path (`float`, defaults to 0.0): + Stochastic depth rate. + """ + + def __init__(self, config: ConvNextV2Config, dim: int, drop_path: float = 0.0, **kwargs): + super().__init__(**kwargs) + self.dim = dim + self.config = config + self.dwconv = keras.layers.Conv2D( + filters=dim, + kernel_size=7, + padding="same", + groups=dim, + kernel_initializer=get_initializer(config.initializer_range), + bias_initializer=keras.initializers.Zeros(), + name="dwconv", + ) # depthwise conv + self.layernorm = keras.layers.LayerNormalization( + epsilon=1e-6, + name="layernorm", + ) + self.pwconv1 = keras.layers.Dense( + units=4 * dim, + kernel_initializer=get_initializer(config.initializer_range), + bias_initializer=keras.initializers.Zeros(), + name="pwconv1", + ) # pointwise/1x1 convs, implemented with linear layers + self.act = get_tf_activation(config.hidden_act) + self.grn = TFConvNextV2GRN(config, 4 * dim, dtype=tf.float32, name="grn") + self.pwconv2 = keras.layers.Dense( + units=dim, + kernel_initializer=get_initializer(config.initializer_range), + bias_initializer=keras.initializers.Zeros(), + name="pwconv2", + ) + # Using `layers.Activation` instead of `tf.identity` to better control `training` + # behaviour. + self.drop_path = ( + TFConvNextV2DropPath(drop_path, name="drop_path") + if drop_path > 0.0 + else keras.layers.Activation("linear", name="drop_path") + ) + + def call(self, hidden_states, training=False): + input = hidden_states + x = self.dwconv(hidden_states) + x = self.layernorm(x) + x = self.pwconv1(x) + x = self.act(x) + x = self.grn(x) + x = self.pwconv2(x) + x = self.drop_path(x, training=training) + x = input + x + return x + + def build(self, input_shape=None): + if self.built: + return + self.built = True + if getattr(self, "dwconv", None) is not None: + with tf.name_scope(self.dwconv.name): + self.dwconv.build([None, None, None, self.dim]) + if getattr(self, "layernorm", None) is not None: + with tf.name_scope(self.layernorm.name): + self.layernorm.build([None, None, None, self.dim]) + if getattr(self, "pwconv1", None) is not None: + with tf.name_scope(self.pwconv1.name): + self.pwconv1.build([None, None, self.dim]) + if getattr(self, "grn", None) is not None: + with tf.name_scope(self.grn.name): + self.grn.build(None) + if getattr(self, "pwconv2", None) is not None: + with tf.name_scope(self.pwconv2.name): + self.pwconv2.build([None, None, 4 * self.dim]) + if getattr(self, "drop_path", None) is not None: + with tf.name_scope(self.drop_path.name): + self.drop_path.build(None) + + +# Copied from transformers.models.convnext.modeling_tf_convnext.TFConvNextStage with ConvNext->ConvNextV2 +class TFConvNextV2Stage(keras.layers.Layer): + """ConvNextV2 stage, consisting of an optional downsampling layer + multiple residual blocks. + + Args: + config (`ConvNextV2V2Config`): + Model configuration class. + in_channels (`int`): + Number of input channels. + out_channels (`int`): + Number of output channels. + depth (`int`): + Number of residual blocks. + drop_path_rates(`List[float]`): + Stochastic depth rates for each layer. + """ + + def __init__( + self, + config: ConvNextV2Config, + in_channels: int, + out_channels: int, + kernel_size: int = 2, + stride: int = 2, + depth: int = 2, + drop_path_rates: Optional[List[float]] = None, + **kwargs, + ): + super().__init__(**kwargs) + if in_channels != out_channels or stride > 1: + self.downsampling_layer = [ + keras.layers.LayerNormalization( + epsilon=1e-6, + name="downsampling_layer.0", + ), + # Inputs to this layer will follow NHWC format since we + # transposed the inputs from NCHW to NHWC in the `TFConvNextV2Embeddings` + # layer. All the outputs throughout the model will be in NHWC + # from this point on until the output where we again change to + # NCHW. + keras.layers.Conv2D( + filters=out_channels, + kernel_size=kernel_size, + strides=stride, + kernel_initializer=get_initializer(config.initializer_range), + bias_initializer=keras.initializers.Zeros(), + name="downsampling_layer.1", + ), + ] + else: + self.downsampling_layer = [tf.identity] + + drop_path_rates = drop_path_rates or [0.0] * depth + self.layers = [ + TFConvNextV2Layer( + config, + dim=out_channels, + drop_path=drop_path_rates[j], + name=f"layers.{j}", + ) + for j in range(depth) + ] + self.in_channels = in_channels + self.out_channels = out_channels + self.stride = stride + + def call(self, hidden_states): + for layer in self.downsampling_layer: + hidden_states = layer(hidden_states) + for layer in self.layers: + hidden_states = layer(hidden_states) + return hidden_states + + def build(self, input_shape=None): + if self.built: + return + self.built = True + if getattr(self, "layers", None) is not None: + for layer in self.layers: + with tf.name_scope(layer.name): + layer.build(None) + if self.in_channels != self.out_channels or self.stride > 1: + with tf.name_scope(self.downsampling_layer[0].name): + self.downsampling_layer[0].build([None, None, None, self.in_channels]) + with tf.name_scope(self.downsampling_layer[1].name): + self.downsampling_layer[1].build([None, None, None, self.in_channels]) + + +class TFConvNextV2Encoder(keras.layers.Layer): + def __init__(self, config: ConvNextV2Config, **kwargs): + super().__init__(**kwargs) + self.stages = [] + drop_path_rates = tf.linspace(0.0, config.drop_path_rate, sum(config.depths)) + drop_path_rates = tf.split(drop_path_rates, config.depths) + drop_path_rates = [x.numpy().tolist() for x in drop_path_rates] + prev_chs = config.hidden_sizes[0] + for i in range(config.num_stages): + out_chs = config.hidden_sizes[i] + stage = TFConvNextV2Stage( + config, + in_channels=prev_chs, + out_channels=out_chs, + stride=2 if i > 0 else 1, + depth=config.depths[i], + drop_path_rates=drop_path_rates[i], + name=f"stages.{i}", + ) + self.stages.append(stage) + prev_chs = out_chs + + def call( + self, + hidden_states: tf.Tensor, + output_hidden_states: Optional[bool] = False, + return_dict: Optional[bool] = True, + ) -> Union[Tuple, TFBaseModelOutputWithNoAttention]: + all_hidden_states = () if output_hidden_states else None + + for i, layer_module in enumerate(self.stages): + if output_hidden_states: + all_hidden_states = all_hidden_states + (hidden_states,) + + hidden_states = layer_module(hidden_states) + + if output_hidden_states: + all_hidden_states = all_hidden_states + (hidden_states,) + + if not return_dict: + return tuple(v for v in [hidden_states, all_hidden_states] if v is not None) + + return TFBaseModelOutputWithNoAttention(last_hidden_state=hidden_states, hidden_states=all_hidden_states) + + def build(self, input_shape=None): + for stage in self.stages: + with tf.name_scope(stage.name): + stage.build(None) + + +@keras_serializable +class TFConvNextV2MainLayer(keras.layers.Layer): + config_class = ConvNextV2Config + + def __init__(self, config: ConvNextV2Config, **kwargs): + super().__init__(**kwargs) + + self.config = config + self.embeddings = TFConvNextV2Embeddings(config, name="embeddings") + self.encoder = TFConvNextV2Encoder(config, name="encoder") + self.layernorm = keras.layers.LayerNormalization(epsilon=config.layer_norm_eps, name="layernorm") + # We are setting the `data_format` like so because from here on we will revert to the + # NCHW output format + self.pooler = keras.layers.GlobalAvgPool2D(data_format="channels_last") + + @unpack_inputs + def call( + self, + pixel_values: TFModelInputType | None = None, + output_hidden_states: Optional[bool] = None, + return_dict: Optional[bool] = None, + training: bool = False, + ) -> Union[TFBaseModelOutputWithPooling, Tuple[tf.Tensor]]: + output_hidden_states = ( + output_hidden_states if output_hidden_states is not None else self.config.output_hidden_states + ) + return_dict = return_dict if return_dict is not None else self.config.use_return_dict + + if pixel_values is None: + raise ValueError("You have to specify pixel_values") + + embedding_output = self.embeddings(pixel_values, training=training) + + encoder_outputs = self.encoder( + embedding_output, + output_hidden_states=output_hidden_states, + return_dict=return_dict, + training=training, + ) + + last_hidden_state = encoder_outputs[0] + + # Change to NCHW output format have uniformity in the modules + pooled_output = self.pooler(last_hidden_state) + last_hidden_state = tf.transpose(last_hidden_state, perm=(0, 3, 1, 2)) + pooled_output = self.layernorm(pooled_output) + + # Change the other hidden state outputs to NCHW as well + if output_hidden_states: + hidden_states = tuple([tf.transpose(h, perm=(0, 3, 1, 2)) for h in encoder_outputs[1]]) + + if not return_dict: + hidden_states = hidden_states if output_hidden_states else () + return (last_hidden_state, pooled_output) + hidden_states + + return TFBaseModelOutputWithPoolingAndNoAttention( + last_hidden_state=last_hidden_state, + pooler_output=pooled_output, + hidden_states=hidden_states if output_hidden_states else encoder_outputs.hidden_states, + ) + + def build(self, input_shape=None): + if self.built: + return + self.built = True + if getattr(self, "embeddings", None) is not None: + with tf.name_scope(self.embeddings.name): + self.embeddings.build(None) + if getattr(self, "encoder", None) is not None: + with tf.name_scope(self.encoder.name): + self.encoder.build(None) + if getattr(self, "layernorm", None) is not None: + with tf.name_scope(self.layernorm.name): + self.layernorm.build([None, self.config.hidden_sizes[-1]]) + + +class TFConvNextV2PreTrainedModel(TFPreTrainedModel): + """ + An abstract class to handle weights initialization and a simple interface for downloading and loading pretrained + models. + """ + + config_class = ConvNextV2Config + base_model_prefix = "convnextv2" + main_input_name = "pixel_values" + + +CONVNEXTV2_START_DOCSTRING = r""" + This model inherits from [`TFPreTrainedModel`]. Check the superclass documentation for the generic methods the + library implements for all its model (such as downloading or saving, resizing the input embeddings, pruning heads + etc.) + + This model is also a [keras.Model](https://www.tensorflow.org/api_docs/python/tf/keras/Model) subclass. Use it + as a regular TF 2.0 Keras Model and refer to the TF 2.0 documentation for all matter related to general usage and + behavior. + + + + TensorFlow models and layers in `transformers` accept two formats as input: + + - having all inputs as keyword arguments (like PyTorch models), or + - having all inputs as a list, tuple or dict in the first positional argument. + + The reason the second format is supported is that Keras methods prefer this format when passing inputs to models + and layers. Because of this support, when using methods like `model.fit()` things should "just work" for you - just + pass your inputs and labels in any format that `model.fit()` supports! If, however, you want to use the second + format outside of Keras methods like `fit()` and `predict()`, such as when creating your own layers or models with + the Keras `Functional` API, there are three possibilities you can use to gather all the input Tensors in the first + positional argument: + + - a single Tensor with `pixel_values` only and nothing else: `model(pixel_values)` + - a list of varying length with one or several input Tensors IN THE ORDER given in the docstring: + `model([pixel_values, attention_mask])` or `model([pixel_values, attention_mask, token_type_ids])` + - a dictionary with one or several input Tensors associated to the input names given in the docstring: + `model({"pixel_values": pixel_values, "token_type_ids": token_type_ids})` + + Note that when creating models and layers with + [subclassing](https://keras.io/guides/making_new_layers_and_models_via_subclassing/) then you don't need to worry + about any of this, as you can just pass inputs like you would to any other Python function! + + + + Parameters: + config ([`ConvNextV2Config`]): Model configuration class with all the parameters of the model. + Initializing with a config file does not load the weights associated with the model, only the + configuration. Check out the [`~TFPreTrainedModel.from_pretrained`] method to load the model weights. +""" + +CONVNEXTV2_INPUTS_DOCSTRING = r""" + Args: + pixel_values (`np.ndarray`, `tf.Tensor`, `List[tf.Tensor]`, `Dict[str, tf.Tensor]` or `Dict[str, np.ndarray]` and each example must have the shape `(batch_size, num_channels, height, width)`): + Pixel values. Pixel values can be obtained using [`AutoImageProcessor`]. See + [`ConvNextImageProcessor.__call__`] for details. + + output_hidden_states (`bool`, *optional*): + Whether or not to return the hidden states of all layers. See `hidden_states` under returned tensors for + more detail. This argument can be used only in eager mode, in graph mode the value in the config will be + used instead. + return_dict (`bool`, *optional*): + Whether or not to return a [`~utils.ModelOutput`] instead of a plain tuple. This argument can be used in + eager mode, in graph mode the value will always be set to `True`. +""" + + +@add_start_docstrings( + "The bare ConvNextV2 model outputting raw features without any specific head on top.", + CONVNEXTV2_START_DOCSTRING, +) +class TFConvNextV2Model(TFConvNextV2PreTrainedModel): + def __init__(self, config: ConvNextV2Config, *inputs, **kwargs): + super().__init__(config, *inputs, **kwargs) + self.convnextv2 = TFConvNextV2MainLayer(config, name="convnextv2") + + @unpack_inputs + @add_start_docstrings_to_model_forward(CONVNEXTV2_INPUTS_DOCSTRING) + @add_code_sample_docstrings( + checkpoint=_CHECKPOINT_FOR_DOC, + output_type=TFBaseModelOutputWithPoolingAndNoAttention, + config_class=_CONFIG_FOR_DOC, + modality="vision", + expected_output=_EXPECTED_OUTPUT_SHAPE, + ) + def call( + self, + pixel_values: TFModelInputType | None = None, + output_hidden_states: Optional[bool] = None, + return_dict: Optional[bool] = None, + training: bool = False, + ) -> Union[TFBaseModelOutputWithPoolingAndNoAttention, Tuple[tf.Tensor]]: + output_hidden_states = ( + output_hidden_states if output_hidden_states is not None else self.config.output_hidden_states + ) + return_dict = return_dict if return_dict is not None else self.config.use_return_dict + + if pixel_values is None: + raise ValueError("You have to specify pixel_values") + + outputs = self.convnextv2( + pixel_values=pixel_values, + output_hidden_states=output_hidden_states, + return_dict=return_dict, + training=training, + ) + + if not return_dict: + return outputs[:] + + return TFBaseModelOutputWithPoolingAndNoAttention( + last_hidden_state=outputs.last_hidden_state, + pooler_output=outputs.pooler_output, + hidden_states=outputs.hidden_states, + ) + + def build(self, input_shape=None): + if self.built: + return + self.built = True + if getattr(self, "convnextv2", None) is not None: + with tf.name_scope(self.convnextv2.name): + self.convnextv2.build(None) + + +@add_start_docstrings( + """ + ConvNextV2 Model with an image classification head on top (a linear layer on top of the pooled features), e.g. for + ImageNet. + """, + CONVNEXTV2_START_DOCSTRING, +) +class TFConvNextV2ForImageClassification(TFConvNextV2PreTrainedModel, TFSequenceClassificationLoss): + def __init__(self, config: ConvNextV2Config, *inputs, **kwargs): + super().__init__(config, *inputs, **kwargs) + + self.num_labels = config.num_labels + self.convnextv2 = TFConvNextV2MainLayer(config, name="convnextv2") + + # Classifier head + self.classifier = keras.layers.Dense( + units=config.num_labels, + kernel_initializer=get_initializer(config.initializer_range), + bias_initializer=keras.initializers.Zeros(), + name="classifier", + ) + + @unpack_inputs + @add_start_docstrings_to_model_forward(CONVNEXTV2_INPUTS_DOCSTRING) + @add_code_sample_docstrings( + checkpoint=_IMAGE_CLASS_CHECKPOINT, + output_type=TFImageClassifierOutputWithNoAttention, + config_class=_CONFIG_FOR_DOC, + expected_output=_IMAGE_CLASS_EXPECTED_OUTPUT, + ) + def call( + self, + pixel_values: TFModelInputType | None = None, + output_hidden_states: Optional[bool] = None, + return_dict: Optional[bool] = None, + labels: np.ndarray | tf.Tensor | None = None, + training: Optional[bool] = False, + ) -> Union[TFImageClassifierOutputWithNoAttention, Tuple[tf.Tensor]]: + r""" + labels (`tf.Tensor` or `np.ndarray` of shape `(batch_size,)`, *optional*): + Labels for computing the image classification/regression loss. Indices should be in `[0, ..., + config.num_labels - 1]`. If `config.num_labels == 1` a regression loss is computed (Mean-Square loss), If + `config.num_labels > 1` a classification loss is computed (Cross-Entropy). + """ + output_hidden_states = ( + output_hidden_states if output_hidden_states is not None else self.config.output_hidden_states + ) + return_dict = return_dict if return_dict is not None else self.config.use_return_dict + + if pixel_values is None: + raise ValueError("You have to specify pixel_values") + + outputs = self.convnextv2( + pixel_values, + output_hidden_states=output_hidden_states, + return_dict=return_dict, + training=training, + ) + + pooled_output = outputs.pooler_output if return_dict else outputs[1] + + logits = self.classifier(pooled_output) + loss = None if labels is None else self.hf_compute_loss(labels=labels, logits=logits) + + if not return_dict: + output = (logits,) + outputs[2:] + return ((loss,) + output) if loss is not None else output + + return TFImageClassifierOutputWithNoAttention( + loss=loss, + logits=logits, + hidden_states=outputs.hidden_states, + ) + + def build(self, input_shape=None): + if self.built: + return + self.built = True + if getattr(self, "convnextv2", None) is not None: + with tf.name_scope(self.convnextv2.name): + self.convnextv2.build(None) + if getattr(self, "classifier", None) is not None: + with tf.name_scope(self.classifier.name): + self.classifier.build([None, None, self.config.hidden_sizes[-1]]) diff --git a/venv/lib/python3.10/site-packages/transformers/models/fastspeech2_conformer/tokenization_fastspeech2_conformer.py b/venv/lib/python3.10/site-packages/transformers/models/fastspeech2_conformer/tokenization_fastspeech2_conformer.py new file mode 100644 index 0000000000000000000000000000000000000000..5b979c8761c42cfffbb215578430d82596b961fc --- /dev/null +++ b/venv/lib/python3.10/site-packages/transformers/models/fastspeech2_conformer/tokenization_fastspeech2_conformer.py @@ -0,0 +1,184 @@ +# coding=utf-8 +# Copyright 2023 The HuggingFace Team and The HuggingFace Inc. team. All rights reserved. +# +# Licensed under the Apache License, Version 2.0 (the "License"); +# you may not use this file except in compliance with the License. +# You may obtain a copy of the License at +# +# http://www.apache.org/licenses/LICENSE-2.0 +# +# Unless required by applicable law or agreed to in writing, software +# distributed under the License is distributed on an "AS IS" BASIS, +# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. +# See the License for the specific language governing permissions and +# limitations under the License. +"""Tokenization classes for FastSpeech2Conformer.""" +import json +import os +from typing import Optional, Tuple + +import regex + +from ...tokenization_utils import PreTrainedTokenizer +from ...utils import logging, requires_backends + + +logger = logging.get_logger(__name__) + +VOCAB_FILES_NAMES = {"vocab_file": "vocab.json"} + + +class FastSpeech2ConformerTokenizer(PreTrainedTokenizer): + """ + Construct a FastSpeech2Conformer tokenizer. + + Args: + vocab_file (`str`): + Path to the vocabulary file. + bos_token (`str`, *optional*, defaults to `""`): + The begin of sequence token. Note that for FastSpeech2, it is the same as the `eos_token`. + eos_token (`str`, *optional*, defaults to `""`): + The end of sequence token. Note that for FastSpeech2, it is the same as the `bos_token`. + pad_token (`str`, *optional*, defaults to `""`): + The token used for padding, for example when batching sequences of different lengths. + unk_token (`str`, *optional*, defaults to `""`): + The unknown token. A token that is not in the vocabulary cannot be converted to an ID and is set to be this + token instead. + should_strip_spaces (`bool`, *optional*, defaults to `False`): + Whether or not to strip the spaces from the list of tokens. + """ + + vocab_files_names = VOCAB_FILES_NAMES + model_input_names = ["input_ids", "attention_mask"] + + def __init__( + self, + vocab_file, + bos_token="", + eos_token="", + pad_token="", + unk_token="", + should_strip_spaces=False, + **kwargs, + ): + requires_backends(self, "g2p_en") + + with open(vocab_file, encoding="utf-8") as vocab_handle: + self.encoder = json.load(vocab_handle) + + import g2p_en + + self.g2p = g2p_en.G2p() + + self.decoder = {v: k for k, v in self.encoder.items()} + + super().__init__( + bos_token=bos_token, + eos_token=eos_token, + unk_token=unk_token, + pad_token=pad_token, + should_strip_spaces=should_strip_spaces, + **kwargs, + ) + + self.should_strip_spaces = should_strip_spaces + + @property + def vocab_size(self): + return len(self.decoder) + + def get_vocab(self): + "Returns vocab as a dict" + return dict(self.encoder, **self.added_tokens_encoder) + + def prepare_for_tokenization(self, text, is_split_into_words=False, **kwargs): + # expand symbols + text = regex.sub(";", ",", text) + text = regex.sub(":", ",", text) + text = regex.sub("-", " ", text) + text = regex.sub("&", "and", text) + + # strip unnecessary symbols + text = regex.sub(r"[\(\)\[\]\<\>\"]+", "", text) + + # strip whitespaces + text = regex.sub(r"\s+", " ", text) + + text = text.upper() + + return text, kwargs + + def _tokenize(self, text): + """Returns a tokenized string.""" + # phonemize + tokens = self.g2p(text) + + if self.should_strip_spaces: + tokens = list(filter(lambda s: s != " ", tokens)) + + tokens.append(self.eos_token) + + return tokens + + def _convert_token_to_id(self, token): + """Converts a token (str) in an id using the vocab.""" + return self.encoder.get(token, self.encoder.get(self.unk_token)) + + def _convert_id_to_token(self, index): + """Converts an index (integer) in a token (str) using the vocab.""" + return self.decoder.get(index, self.unk_token) + + # Override since phonemes cannot be converted back to strings + def decode(self, token_ids, **kwargs): + logger.warning( + "Phonemes cannot be reliably converted to a string due to the one-many mapping, converting to tokens instead." + ) + return self.convert_ids_to_tokens(token_ids) + + # Override since phonemes cannot be converted back to strings + def convert_tokens_to_string(self, tokens, **kwargs): + logger.warning( + "Phonemes cannot be reliably converted to a string due to the one-many mapping, returning the tokens." + ) + return tokens + + def save_vocabulary(self, save_directory: str, filename_prefix: Optional[str] = None) -> Tuple[str]: + """ + Save the vocabulary and special tokens file to a directory. + + Args: + save_directory (`str`): + The directory in which to save the vocabulary. + + Returns: + `Tuple(str)`: Paths to the files saved. + """ + if not os.path.isdir(save_directory): + logger.error(f"Vocabulary path ({save_directory}) should be a directory") + return + vocab_file = os.path.join( + save_directory, (filename_prefix + "-" if filename_prefix else "") + VOCAB_FILES_NAMES["vocab_file"] + ) + + with open(vocab_file, "w", encoding="utf-8") as f: + f.write(json.dumps(self.get_vocab(), ensure_ascii=False)) + + return (vocab_file,) + + def __getstate__(self): + state = self.__dict__.copy() + state["g2p"] = None + return state + + def __setstate__(self, d): + self.__dict__ = d + + try: + import g2p_en + + self.g2p = g2p_en.G2p() + except ImportError: + raise ImportError( + "You need to install g2p-en to use FastSpeech2ConformerTokenizer. " + "See https://pypi.org/project/g2p-en/ for installation." + ) diff --git a/venv/lib/python3.10/site-packages/transformers/models/mobilenet_v1/__pycache__/convert_original_tf_checkpoint_to_pytorch.cpython-310.pyc b/venv/lib/python3.10/site-packages/transformers/models/mobilenet_v1/__pycache__/convert_original_tf_checkpoint_to_pytorch.cpython-310.pyc new file mode 100644 index 0000000000000000000000000000000000000000..87e1162ddff8d6d739299fb0b628830b0a1207aa Binary files /dev/null and b/venv/lib/python3.10/site-packages/transformers/models/mobilenet_v1/__pycache__/convert_original_tf_checkpoint_to_pytorch.cpython-310.pyc differ diff --git a/venv/lib/python3.10/site-packages/transformers/models/mobilenet_v1/__pycache__/modeling_mobilenet_v1.cpython-310.pyc b/venv/lib/python3.10/site-packages/transformers/models/mobilenet_v1/__pycache__/modeling_mobilenet_v1.cpython-310.pyc new file mode 100644 index 0000000000000000000000000000000000000000..cb08c8f9d1b207a973a54c05fd4fb03a73952212 Binary files /dev/null and b/venv/lib/python3.10/site-packages/transformers/models/mobilenet_v1/__pycache__/modeling_mobilenet_v1.cpython-310.pyc differ diff --git a/venv/lib/python3.10/site-packages/transformers/models/prophetnet/__init__.py b/venv/lib/python3.10/site-packages/transformers/models/prophetnet/__init__.py new file mode 100644 index 0000000000000000000000000000000000000000..083301cc20c677fa15daa9cff63385f04fcd0507 --- /dev/null +++ b/venv/lib/python3.10/site-packages/transformers/models/prophetnet/__init__.py @@ -0,0 +1,65 @@ +# Copyright 2020 The HuggingFace Team. All rights reserved. +# +# Licensed under the Apache License, Version 2.0 (the "License"); +# you may not use this file except in compliance with the License. +# You may obtain a copy of the License at +# +# http://www.apache.org/licenses/LICENSE-2.0 +# +# Unless required by applicable law or agreed to in writing, software +# distributed under the License is distributed on an "AS IS" BASIS, +# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. +# See the License for the specific language governing permissions and +# limitations under the License. + +from typing import TYPE_CHECKING + +from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_torch_available + + +_import_structure = { + "configuration_prophetnet": ["PROPHETNET_PRETRAINED_CONFIG_ARCHIVE_MAP", "ProphetNetConfig"], + "tokenization_prophetnet": ["ProphetNetTokenizer"], +} + +try: + if not is_torch_available(): + raise OptionalDependencyNotAvailable() +except OptionalDependencyNotAvailable: + pass +else: + _import_structure["modeling_prophetnet"] = [ + "PROPHETNET_PRETRAINED_MODEL_ARCHIVE_LIST", + "ProphetNetDecoder", + "ProphetNetEncoder", + "ProphetNetForCausalLM", + "ProphetNetForConditionalGeneration", + "ProphetNetModel", + "ProphetNetPreTrainedModel", + ] + + +if TYPE_CHECKING: + from .configuration_prophetnet import PROPHETNET_PRETRAINED_CONFIG_ARCHIVE_MAP, ProphetNetConfig + from .tokenization_prophetnet import ProphetNetTokenizer + + try: + if not is_torch_available(): + raise OptionalDependencyNotAvailable() + except OptionalDependencyNotAvailable: + pass + else: + from .modeling_prophetnet import ( + PROPHETNET_PRETRAINED_MODEL_ARCHIVE_LIST, + ProphetNetDecoder, + ProphetNetEncoder, + ProphetNetForCausalLM, + ProphetNetForConditionalGeneration, + ProphetNetModel, + ProphetNetPreTrainedModel, + ) + +else: + import sys + + sys.modules[__name__] = _LazyModule(__name__, globals()["__file__"], _import_structure, module_spec=__spec__) diff --git a/venv/lib/python3.10/site-packages/transformers/models/prophetnet/__pycache__/__init__.cpython-310.pyc b/venv/lib/python3.10/site-packages/transformers/models/prophetnet/__pycache__/__init__.cpython-310.pyc new file mode 100644 index 0000000000000000000000000000000000000000..bddf31dbdf94151c7e8bfa3e767c0bebc4f7706c Binary files /dev/null and b/venv/lib/python3.10/site-packages/transformers/models/prophetnet/__pycache__/__init__.cpython-310.pyc differ diff --git a/venv/lib/python3.10/site-packages/transformers/models/prophetnet/__pycache__/configuration_prophetnet.cpython-310.pyc b/venv/lib/python3.10/site-packages/transformers/models/prophetnet/__pycache__/configuration_prophetnet.cpython-310.pyc new file mode 100644 index 0000000000000000000000000000000000000000..dc617f319c1f419258d381023a43b8a86deb9988 Binary files /dev/null and b/venv/lib/python3.10/site-packages/transformers/models/prophetnet/__pycache__/configuration_prophetnet.cpython-310.pyc differ diff --git a/venv/lib/python3.10/site-packages/transformers/models/prophetnet/__pycache__/convert_prophetnet_original_pytorch_checkpoint_to_pytorch.cpython-310.pyc b/venv/lib/python3.10/site-packages/transformers/models/prophetnet/__pycache__/convert_prophetnet_original_pytorch_checkpoint_to_pytorch.cpython-310.pyc new file mode 100644 index 0000000000000000000000000000000000000000..91cd9aeed3b752547fafa2f7281dad432375f687 Binary files /dev/null and b/venv/lib/python3.10/site-packages/transformers/models/prophetnet/__pycache__/convert_prophetnet_original_pytorch_checkpoint_to_pytorch.cpython-310.pyc differ diff --git a/venv/lib/python3.10/site-packages/transformers/models/prophetnet/__pycache__/modeling_prophetnet.cpython-310.pyc b/venv/lib/python3.10/site-packages/transformers/models/prophetnet/__pycache__/modeling_prophetnet.cpython-310.pyc new file mode 100644 index 0000000000000000000000000000000000000000..92f7e80c21077e30d784022c3c5c20b8b0342e30 Binary files /dev/null and b/venv/lib/python3.10/site-packages/transformers/models/prophetnet/__pycache__/modeling_prophetnet.cpython-310.pyc differ diff --git a/venv/lib/python3.10/site-packages/transformers/models/prophetnet/__pycache__/tokenization_prophetnet.cpython-310.pyc b/venv/lib/python3.10/site-packages/transformers/models/prophetnet/__pycache__/tokenization_prophetnet.cpython-310.pyc new file mode 100644 index 0000000000000000000000000000000000000000..3d912a55be7821effcb6a5e782332408d90f4935 Binary files /dev/null and b/venv/lib/python3.10/site-packages/transformers/models/prophetnet/__pycache__/tokenization_prophetnet.cpython-310.pyc differ diff --git a/venv/lib/python3.10/site-packages/transformers/models/prophetnet/configuration_prophetnet.py b/venv/lib/python3.10/site-packages/transformers/models/prophetnet/configuration_prophetnet.py new file mode 100644 index 0000000000000000000000000000000000000000..e07936a14cd30245a28b7e1619d1b46ac0ca9f63 --- /dev/null +++ b/venv/lib/python3.10/site-packages/transformers/models/prophetnet/configuration_prophetnet.py @@ -0,0 +1,180 @@ +# coding=utf-8 +# Copyright 2020 The Microsoft Authors and The HuggingFace Inc. team. +# +# Licensed under the Apache License, Version 2.0 (the "License"); +# you may not use this file except in compliance with the License. +# You may obtain a copy of the License at +# +# http://www.apache.org/licenses/LICENSE-2.0 +# +# Unless required by applicable law or agreed to in writing, software +# distributed under the License is distributed on an "AS IS" BASIS, +# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. +# See the License for the specific language governing permissions and +# limitations under the License. +""" ProphetNet model configuration""" + +from typing import Callable, Optional, Union + +from ...configuration_utils import PretrainedConfig +from ...utils import logging + + +logger = logging.get_logger(__name__) + + +from ..deprecated._archive_maps import PROPHETNET_PRETRAINED_CONFIG_ARCHIVE_MAP # noqa: F401, E402 + + +class ProphetNetConfig(PretrainedConfig): + r""" + This is the configuration class to store the configuration of a [`ProphetNetModel`]. It is used to instantiate a + ProphetNet model according to the specified arguments, defining the model architecture. Instantiating a + configuration with the defaults will yield a similar configuration to that of the ProphetNet + [microsoft/prophetnet-large-uncased](https://huggingface.co/microsoft/prophetnet-large-uncased) architecture. + + Configuration objects inherit from [`PretrainedConfig`] and can be used to control the model outputs. Read the + documentation from [`PretrainedConfig`] for more information. + + Args: + activation_dropout (`float`, *optional*, defaults to 0.1): + The dropout ratio for activations inside the fully connected layer. + activation_function (`str` or `function`, *optional*, defaults to `"gelu"`): + The non-linear activation function (function or string) in the encoder and pooler. If string, `"gelu"`, + `"relu"`, `"silu"` and `"gelu_new"` are supported. + vocab_size (`int`, *optional*, defaults to 30522): + Vocabulary size of the ProphetNET model. Defines the number of different tokens that can be represented by + the `inputs_ids` passed when calling [`ProphetNetModel`]. + hidden_size (`int`, *optional*, defaults to 1024): + Dimensionality of the layers and the pooler layer. + encoder_ffn_dim (`int`, *optional*, defaults to 4096): + Dimensionality of the "intermediate" (often named feed-forward) layer in decoder. + num_encoder_layers (`int`, *optional*, defaults to 12): + Number of encoder layers. + num_encoder_attention_heads (`int`, *optional*, defaults to 16): + Number of attention heads for each attention layer in the Transformer encoder. + decoder_ffn_dim (`int`, *optional*, defaults to 4096): + Dimensionality of the `intermediate` (often named feed-forward) layer in decoder. + num_decoder_layers (`int`, *optional*, defaults to 12): + Number of decoder layers. + num_decoder_attention_heads (`int`, *optional*, defaults to 16): + Number of attention heads for each attention layer in the Transformer decoder. + attention_dropout (`float`, *optional*, defaults to 0.1): + The dropout ratio for the attention probabilities. + dropout (`float`, *optional*, defaults to 0.1): + The dropout probability for all fully connected layers in the embeddings, encoder, and pooler. + max_position_embeddings (`int`, *optional*, defaults to 512): + The maximum sequence length that this model might ever be used with. Typically set this to something large + just in case (e.g., 512 or 1024 or 2048). + init_std (`float`, *optional*, defaults to 0.02): + The standard deviation of the truncated_normal_initializer for initializing all weight matrices. + add_cross_attention (`bool`, *optional*, defaults to `True`): + Whether cross-attention layers should be added to the model. + is_encoder_decoder (`bool`, *optional*, defaults to `True`): + Whether this is an encoder/decoder model. + pad_token_id (`int`, *optional*, defaults to 1) + Padding token id. + bos_token_id (`int`, *optional*, defaults to 0) + Beginning of stream token id. + eos_token_id (`int`, *optional*, defaults to 2) + End of stream token id. + ngram (`int`, *optional*, defaults to 2) + Number of future tokens to predict. Set to 1 to be same as traditional Language model to predict next first + token. + num_buckets (`int`, *optional*, defaults to 32) + The number of buckets to use for each attention layer. This is for relative position calculation. See the + [T5 paper](see https://arxiv.org/abs/1910.10683) for more details. + relative_max_distance (`int`, *optional*, defaults to 128) + Relative distances greater than this number will be put into the last same bucket. This is for relative + position calculation. See the [T5 paper](see https://arxiv.org/abs/1910.10683) for more details. + disable_ngram_loss (`bool`, *optional*, defaults to `False`): + Whether be trained predicting only the next first token. + eps (`float`, *optional*, defaults to 0.0): + Controls the `epsilon` parameter value for label smoothing in the loss calculation. If set to 0, no label + smoothing is performed. + use_cache (`bool`, *optional*, defaults to `True`): + Whether or not the model should return the last key/values attentions (not used by all models). + """ + + model_type = "prophetnet" + keys_to_ignore_at_inference = ["past_key_values"] + attribute_map = { + "num_attention_heads": "num_encoder_attention_heads", + } + + def __init__( + self, + activation_dropout: Optional[float] = 0.1, + activation_function: Optional[Union[str, Callable]] = "gelu", + vocab_size: Optional[int] = 30522, + hidden_size: Optional[int] = 1024, + encoder_ffn_dim: Optional[int] = 4096, + num_encoder_layers: Optional[int] = 12, + num_encoder_attention_heads: Optional[int] = 16, + decoder_ffn_dim: Optional[int] = 4096, + num_decoder_layers: Optional[int] = 12, + num_decoder_attention_heads: Optional[int] = 16, + attention_dropout: Optional[float] = 0.1, + dropout: Optional[float] = 0.1, + max_position_embeddings: Optional[int] = 512, + init_std: Optional[float] = 0.02, + is_encoder_decoder: Optional[bool] = True, + add_cross_attention: Optional[bool] = True, + decoder_start_token_id: Optional[int] = 0, + ngram: Optional[int] = 2, + num_buckets: Optional[int] = 32, + relative_max_distance: Optional[int] = 128, + disable_ngram_loss: Optional[bool] = False, + eps: Optional[float] = 0.0, + use_cache: Optional[bool] = True, + pad_token_id: Optional[int] = 0, + bos_token_id: Optional[int] = 1, + eos_token_id: Optional[int] = 2, + **kwargs, + ): + self.vocab_size = vocab_size + self.hidden_size = hidden_size + self.encoder_ffn_dim = encoder_ffn_dim + self.num_encoder_layers = num_encoder_layers + self.num_encoder_attention_heads = num_encoder_attention_heads + self.decoder_ffn_dim = decoder_ffn_dim + self.num_decoder_layers = num_decoder_layers + self.num_decoder_attention_heads = num_decoder_attention_heads + self.max_position_embeddings = max_position_embeddings + self.init_std = init_std # Normal(0, this parameter) + self.activation_function = activation_function + + # parameters for prophetnet + self.ngram = ngram + self.num_buckets = num_buckets + self.relative_max_distance = relative_max_distance + self.disable_ngram_loss = disable_ngram_loss + self.eps = eps + + # 3 Types of Dropout + self.attention_dropout = attention_dropout + self.activation_dropout = activation_dropout + self.dropout = dropout + + self.use_cache = use_cache + + super().__init__( + pad_token_id=pad_token_id, + bos_token_id=bos_token_id, + eos_token_id=eos_token_id, + is_encoder_decoder=is_encoder_decoder, + add_cross_attention=add_cross_attention, + decoder_start_token_id=decoder_start_token_id, + **kwargs, + ) + + @property + def num_hidden_layers(self) -> int: + return self.num_encoder_layers + self.num_decoder_layers + + @num_hidden_layers.setter + def num_hidden_layers(self, value): + raise NotImplementedError( + "This model does not support the setting of `num_hidden_layers`. Please set `num_encoder_layers` and" + " `num_decoder_layers`." + ) diff --git a/venv/lib/python3.10/site-packages/transformers/models/prophetnet/convert_prophetnet_original_pytorch_checkpoint_to_pytorch.py b/venv/lib/python3.10/site-packages/transformers/models/prophetnet/convert_prophetnet_original_pytorch_checkpoint_to_pytorch.py new file mode 100644 index 0000000000000000000000000000000000000000..c9e64c06ef769ad79fa43e46d6945c9d5f9f86e9 --- /dev/null +++ b/venv/lib/python3.10/site-packages/transformers/models/prophetnet/convert_prophetnet_original_pytorch_checkpoint_to_pytorch.py @@ -0,0 +1,160 @@ +# coding=utf-8 +# Copyright 2020 The HuggingFace Inc. team. +# +# Licensed under the Apache License, Version 2.0 (the "License"); +# you may not use this file except in compliance with the License. +# You may obtain a copy of the License at +# +# http://www.apache.org/licenses/LICENSE-2.0 +# +# Unless required by applicable law or agreed to in writing, software +# distributed under the License is distributed on an "AS IS" BASIS, +# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. +# See the License for the specific language governing permissions and +# limitations under the License. +"""Convert ProphetNet checkpoint.""" + + +import argparse + +from torch import nn + +# transformers_old should correspond to branch `save_old_prophetnet_model_structure` here +# original prophetnet_checkpoints are saved under `patrickvonplaten/..._old` respectively +from transformers_old.modeling_prophetnet import ( + ProphetNetForConditionalGeneration as ProphetNetForConditionalGenerationOld, +) +from transformers_old.modeling_xlm_prophetnet import ( + XLMProphetNetForConditionalGeneration as XLMProphetNetForConditionalGenerationOld, +) + +from transformers import ProphetNetForConditionalGeneration, XLMProphetNetForConditionalGeneration, logging + + +logger = logging.get_logger(__name__) +logging.set_verbosity_info() + + +def convert_prophetnet_checkpoint_to_pytorch(prophetnet_checkpoint_path: str, pytorch_dump_folder_path: str): + """ + Copy/paste/tweak prohpetnet's weights to our prophetnet structure. + """ + if "xprophetnet" in prophetnet_checkpoint_path: + prophet_old = XLMProphetNetForConditionalGenerationOld.from_pretrained(prophetnet_checkpoint_path) + prophet, loading_info = XLMProphetNetForConditionalGeneration.from_pretrained( + prophetnet_checkpoint_path, output_loading_info=True + ) + else: + prophet_old = ProphetNetForConditionalGenerationOld.from_pretrained(prophetnet_checkpoint_path) + prophet, loading_info = ProphetNetForConditionalGeneration.from_pretrained( + prophetnet_checkpoint_path, output_loading_info=True + ) + + special_keys = ["key_proj", "value_proj", "query_proj"] + + mapping = { + "self_attn": "ngram_self_attn", + "cross_attn": "encoder_attn", + "cross_attn_layer_norm": "encoder_attn_layer_norm", + "feed_forward_layer_norm": "final_layer_norm", + "feed_forward": "", + "intermediate": "fc1", + "output": "fc2", + "key_proj": "k_proj", + "query_proj": "q_proj", + "value_proj": "v_proj", + "word_embeddings": "embed_tokens", + "embeddings_layer_norm": "emb_layer_norm", + "relative_pos_embeddings": "relative_linear", + "ngram_embeddings": "ngram_input_embed", + "position_embeddings": "embed_positions", + } + + for key in loading_info["missing_keys"]: + attributes = key.split(".") + + if attributes[0] == "lm_head": + model = prophet + old_model = prophet_old + else: + model = prophet.prophetnet + old_model = prophet_old.model + + is_key_init = False + for attribute in attributes: + if attribute in mapping: + old_attribute = mapping[attribute] + if not hasattr(old_model, old_attribute) and len(old_attribute) > 0: + old_attribute = attribute + elif hasattr(old_model, attribute): + old_attribute = attribute + + if attribute == "weight": + assert old_model.weight.shape == model.weight.shape, "Shapes have to match!" + model.weight = old_model.weight + logger.info(f"{attribute} is initialized.") + is_key_init = True + break + elif attribute == "bias": + assert old_model.bias.shape == model.bias.shape, "Shapes have to match!" + model.bias = old_model.bias + logger.info(f"{attribute} is initialized") + is_key_init = True + break + elif attribute in special_keys and hasattr(old_model, "in_proj_weight"): + embed_dim = old_model.in_proj_weight.shape[0] // 3 + param = getattr(model, attribute) + param.weight.shape == old_model.in_proj_weight[:embed_dim, :].shape, "Shapes have to match" + param.bias.shape == old_model.in_proj_bias[:embed_dim].shape, "Shapes have to match" + if attribute == "query_proj": + model.query_proj.weight = nn.Parameter(old_model.in_proj_weight[:embed_dim, :]) + model.query_proj.bias = nn.Parameter(old_model.in_proj_bias[:embed_dim]) + + elif attribute == "key_proj": + model.key_proj.weight = nn.Parameter(old_model.in_proj_weight[embed_dim : 2 * embed_dim, :]) + model.key_proj.bias = nn.Parameter(old_model.in_proj_bias[embed_dim : 2 * embed_dim]) + elif attribute == "value_proj": + model.value_proj.weight = nn.Parameter(old_model.in_proj_weight[2 * embed_dim :, :]) + model.value_proj.bias = nn.Parameter(old_model.in_proj_bias[2 * embed_dim :]) + is_key_init = True + break + elif attribute == "position_embeddings": + assert ( + model.position_embeddings.weight.shape[-1] == old_model.embed_positions.weight.shape[-1] + ), "Hidden size has to match" + assert model.position_embeddings.weight.shape[0] == 512, "We want 512 position_embeddings." + model.position_embeddings.weight = nn.Parameter(old_model.embed_positions.weight[:512, :]) + is_key_init = True + break + + if attribute.isdigit(): + model = model[int(attribute)] + old_model = old_model[int(old_attribute)] + else: + model = getattr(model, attribute) + + if old_attribute == "": + old_model = old_model + else: + if not hasattr(old_model, old_attribute): + raise ValueError(f"{old_model} does not have {old_attribute}") + old_model = getattr(old_model, old_attribute) + + if not is_key_init: + raise ValueError(f"{key} was not correctly initialized!") + + print(f"Saving model to {pytorch_dump_folder_path}") + prophet.save_pretrained(pytorch_dump_folder_path) + + +if __name__ == "__main__": + parser = argparse.ArgumentParser() + # Required parameters + parser.add_argument( + "--prophetnet_checkpoint_path", default=None, type=str, required=True, help="Path the official PyTorch dump." + ) + parser.add_argument( + "--pytorch_dump_folder_path", default=None, type=str, required=True, help="Path to the output PyTorch model." + ) + args = parser.parse_args() + convert_prophetnet_checkpoint_to_pytorch(args.prophetnet_checkpoint_path, args.pytorch_dump_folder_path) diff --git a/venv/lib/python3.10/site-packages/transformers/models/prophetnet/modeling_prophetnet.py b/venv/lib/python3.10/site-packages/transformers/models/prophetnet/modeling_prophetnet.py new file mode 100644 index 0000000000000000000000000000000000000000..c7d9028cdaf7095a2477fe22d0ccfbd7cff561a1 --- /dev/null +++ b/venv/lib/python3.10/site-packages/transformers/models/prophetnet/modeling_prophetnet.py @@ -0,0 +1,2340 @@ +# coding=utf-8 +# Copyright 2020 The Microsoft Authors and The HuggingFace Inc. team. +# +# Licensed under the Apache License, Version 2.0 (the "License"); +# you may not use this file except in compliance with the License. +# You may obtain a copy of the License at +# +# http://www.apache.org/licenses/LICENSE-2.0 +# +# Unless required by applicable law or agreed to in writing, software +# distributed under the License is distributed on an "AS IS" BASIS, +# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. +# See the License for the specific language governing permissions and +# limitations under the License. +""" PyTorch ProphetNet model, ported from ProphetNet repo(fairsequery_states version).""" + +import copy +import math +import warnings +from dataclasses import dataclass +from typing import Optional, Tuple, Union + +import torch +import torch.utils.checkpoint +from torch import Tensor, nn +from torch.nn import LayerNorm + +from ...activations import ACT2FN +from ...modeling_outputs import BaseModelOutput +from ...modeling_utils import PreTrainedModel +from ...utils import ( + ModelOutput, + add_start_docstrings, + add_start_docstrings_to_model_forward, + logging, + replace_return_docstrings, +) +from .configuration_prophetnet import ProphetNetConfig + + +logger = logging.get_logger(__name__) + +_CONFIG_FOR_DOC = "ProphenetConfig" +_CHECKPOINT_FOR_DOC = "microsoft/prophetnet-large-uncased" + + +from ..deprecated._archive_maps import PROPHETNET_PRETRAINED_MODEL_ARCHIVE_LIST # noqa: F401, E402 + + +PROPHETNET_START_DOCSTRING = r""" + This model inherits from [`PreTrainedModel`]. Check the superclass documentation for the generic methods the + library implements for all its model (such as downloading or saving, resizing the input embeddings, pruning heads + etc.) + + Original ProphetNet code can be found [here](https://github.com/microsoft/ProphetNet). Checkpoints were converted + from original Fairseq checkpoints. For more information on the checkpoint conversion, please take a look at the + file `convert_prophetnet_original_pytorch_checkpoint_to_pytorch.py`. + + This model is a PyTorch [torch.nn.Module](https://pytorch.org/docs/stable/nn.html#torch.nn.Module) sub-class. Use + it as a regular PyTorch Module and refer to the PyTorch documentation for all matters related to general usage and + behavior. + + Parameters: + config ([`ProphetNetConfig`]): Model configuration class with all the parameters of the model. + Initializing with a config file does not load the weights associated with the model, only the + configuration. Check out the [`~PreTrainedModel.from_pretrained`] method to load the model weights. +""" + +PROPHETNET_INPUTS_DOCSTRING = r""" + Args: + input_ids (`torch.LongTensor` of shape `(batch_size, sequence_length)`): + Indices of input sequence tokens in the vocabulary. Padding will be ignored by default should you provide + it. + + Indices can be obtained using [`AutoTokenizer`]. See [`PreTrainedTokenizer.encode`] and + [`PreTrainedTokenizer.__call__`] for details. + + [What are input IDs?](../glossary#input-ids) + attention_mask (`torch.Tensor` of shape `(batch_size, sequence_length)`, *optional*): + Mask to avoid performing attention on padding token indices. Mask values selected in `[0, 1]`: + + - 1 for tokens that are **not masked**, + - 0 for tokens that are **masked**. + + [What are attention masks?](../glossary#attention-mask) + decoder_input_ids (`torch.LongTensor` of shape `(batch_size, target_sequence_length)`, *optional*): + Indices of decoder input sequence tokens in the vocabulary. + + Indices can be obtained using [`AutoTokenizer`]. See [`PreTrainedTokenizer.encode`] and + [`PreTrainedTokenizer.__call__`] for details. + + [What are decoder input IDs?](../glossary#decoder-input-ids) + + ProphetNet uses the `eos_token_id` as the starting token for `decoder_input_ids` generation. If + `past_key_values` is used, optionally only the last `decoder_input_ids` have to be input (see + `past_key_values`). + + decoder_attention_mask (`torch.BoolTensor` of shape `(batch_size, target_sequence_length)`, *optional*): + Default behavior: generate a tensor that ignores pad tokens in `decoder_input_ids`. Causal mask will also + be used by default. + head_mask (`torch.Tensor` of shape `(encoder_layers, encoder_attention_heads)`, *optional*): + Mask to nullify selected heads of the attention modules in the encoder. Mask values selected in `[0, 1]`: + + - 1 indicates the head is **not masked**, + - 0 indicates the head is **masked**. + + decoder_head_mask (`torch.Tensor` of shape `(decoder_layers, decoder_attention_heads)`, *optional*): + Mask to nullify selected heads of the attention modules in the decoder. Mask values selected in `[0, 1]`: + + - 1 indicates the head is **not masked**, + - 0 indicates the head is **masked**. + + cross_attn_head_mask (`torch.Tensor` of shape `(decoder_layers, decoder_attention_heads)`, *optional*): + Mask to nullify selected heads of the cross-attention modules. Mask values selected in `[0, 1]`: + + - 1 indicates the head is **not masked**, + - 0 indicates the head is **masked**. + + encoder_outputs (`tuple(tuple(torch.FloatTensor)`, *optional*): + Tuple consists of (`last_hidden_state`, *optional*: `hidden_states`, *optional*: `attentions`) + `last_hidden_state` of shape `(batch_size, sequence_length, hidden_size)`, *optional*) is a sequence of + hidden-states at the output of the last layer of the encoder. Used in the cross-attention of the decoder. + past_key_values (`tuple(tuple(torch.FloatTensor))` of length `config.n_layers` with each tuple having 4 tensors of shape `(batch_size, num_heads, sequence_length - 1, embed_size_per_head)`): + Contains precomputed key and value hidden-states of the attention blocks. Can be used to speed up decoding. + + If `past_key_values` are used, the user can optionally input only the last `decoder_input_ids` (those that + don't have their past key value states given to this model) of shape `(batch_size, 1)` instead of all + `decoder_input_ids` of shape `(batch_size, sequence_length)`. + use_cache (`bool`, *optional*): + If set to `True`, `past_key_values` key value states are returned and can be used to speed up decoding (see + `past_key_values`). + output_attentions (`bool`, *optional*): + Whether or not to return the attentions tensors of all attention layers. See `attentions` under returned + tensors for more detail. + output_hidden_states (`bool`, *optional*): + Whether or not to return the hidden states of all layers. See `hidden_states` under returned tensors for + more detail. + return_dict (`bool`, *optional*): + Whether or not to return a [`~utils.ModelOutput`] instead of a plain tuple. +""" + +PROPHETNET_STANDALONE_INPUTS_DOCSTRING = r""" + Args: + input_ids (`torch.LongTensor` of shape `(batch_size, sequence_length)`): + Indices of input sequence tokens in the vocabulary. Padding will be ignored by default should you provide + it. + + Indices can be obtained using [`AutoTokenizer`]. See [`PreTrainedTokenizer.encode`] and + [`PreTrainedTokenizer.__call__`] for details. + + [What are input IDs?](../glossary#input-ids) + attention_mask (`torch.Tensor` of shape `(batch_size, sequence_length)`, *optional*): + Mask to avoid performing attention on padding token indices. Mask values selected in `[0, 1]`: + + - 1 for tokens that are **not masked**, + - 0 for tokens that are **masked**. + + [What are attention masks?](../glossary#attention-mask) + head_mask (`torch.Tensor` of shape `(encoder_layers, encoder_attention_heads)`, *optional*): + Mask to nullify selected heads of the attention modules in the encoder. Mask values selected in `[0, 1]`: + + - 1 indicates the head is **not masked**, + - 0 indicates the head is **masked**. + + output_attentions (`bool`, *optional*): + Whether or not to return the attentions tensors of all attention layers. See `attentions` under returned + tensors for more detail. + output_hidden_states (`bool`, *optional*): + Whether or not to return the hidden states of all layers. See `hidden_states` under returned tensors for + more detail. + return_dict (`bool`, *optional*): + Whether or not to return a [`~utils.ModelOutput`] instead of a plain tuple. +""" + + +def softmax(hidden_state, dim, onnx_trace=False): + if onnx_trace: + return nn.functional.softmax(hidden_state.float(), dim=dim) + else: + return nn.functional.softmax(hidden_state, dim=dim, dtype=torch.float32) + + +def ngram_attention_bias(sequence_length, ngram, device, dtype): + """ + This function computes the bias for the predict stream + """ + left_block = ( + torch.ones((ngram, sequence_length, sequence_length), device=device, dtype=dtype) * torch.finfo(dtype).min + ) + right_block = left_block.detach().clone() + # create bias + for stream_idx in range(ngram): + right_block[stream_idx].fill_diagonal_(0, wrap=False) + left_block[stream_idx].triu_(-stream_idx + 1) + + left_block[:, :, 0] = 0 + return torch.cat([left_block, right_block], dim=2) + + +def compute_relative_buckets(num_buckets, max_distance, relative_positions, is_bidirectional=False): + """ + This function computes individual parts of the relative position buckets. For more detail, see paper. + """ + inv_relative_positions = -relative_positions + rel_positions_bucket = 0 + + if is_bidirectional: + num_buckets = num_buckets // 2 + rel_positions_bucket = ( + rel_positions_bucket + + torch.lt(inv_relative_positions, torch.zeros_like(inv_relative_positions)).int() * num_buckets + ) + inv_relative_positions = torch.abs(inv_relative_positions) + else: + inv_relative_positions = torch.max(inv_relative_positions, torch.zeros_like(inv_relative_positions)) + + max_exact = num_buckets // 2 + is_small = torch.lt(inv_relative_positions, max_exact) + val_if_large = max_exact + torch.log(inv_relative_positions.float() / max_exact) / math.log( + max_distance / max_exact + ) * (num_buckets - max_exact) + val_if_large = torch.min(val_if_large, torch.ones_like(val_if_large) * (num_buckets - 1)).int() + rel_positions_bucket = rel_positions_bucket + torch.where(is_small, inv_relative_positions.int(), val_if_large) + return rel_positions_bucket + + +def compute_all_stream_relative_buckets(num_buckets, max_distance, position_ids): + """ + This function computes both main and predict relative position buckets. For more detail, see paper. + """ + # main stream + main_stream_relative_positions = position_ids.unsqueeze(1).repeat(1, position_ids.size(-1), 1) + main_stream_relative_positions = main_stream_relative_positions - position_ids.unsqueeze(-1) + + # predicting stream + predicting_stream_relative_positions = torch.cat((position_ids - 1, position_ids), dim=-1).unsqueeze(1) + predicting_stream_relative_positions = predicting_stream_relative_positions.repeat(1, position_ids.size(-1), 1) + predicting_stream_relative_positions = predicting_stream_relative_positions - position_ids.unsqueeze(-1) + + # get both position buckets + main_relative_position_buckets = compute_relative_buckets( + num_buckets, max_distance, main_stream_relative_positions, is_bidirectional=False + ) + predict_relative_position_buckets = compute_relative_buckets( + num_buckets, max_distance, predicting_stream_relative_positions, is_bidirectional=False + ) + return main_relative_position_buckets, predict_relative_position_buckets + + +@dataclass +class ProphetNetSeq2SeqLMOutput(ModelOutput): + """ + Base class for sequence-to-sequence language models outputs. + + Args: + loss (`torch.FloatTensor` of shape `(1,)`, *optional*, returned when `labels` is provided): + Language modeling loss. + logits (`torch.FloatTensor` of shape `(batch_size, decoder_sequence_length, config.vocab_size)`): + Prediction scores of the main stream language modeling head (scores for each vocabulary token before + SoftMax). + logits_ngram (`torch.FloatTensor` of shape `(batch_size, ngram * decoder_sequence_length, config.vocab_size)`): + Prediction scores of the predict stream language modeling head (scores for each vocabulary token before + SoftMax). + past_key_values (`List[torch.FloatTensor]`, *optional*, returned when `use_cache=True` is passed or when `config.use_cache=True`): + List of `torch.FloatTensor` of length `config.n_layers`, with each tensor of shape `(2, batch_size, + num_attn_heads, decoder_sequence_length, embed_size_per_head)`). + + Contains pre-computed hidden-states (key and values in the attention blocks) of the decoder that can be + used (see `past_key_values` input) to speed up sequential decoding. + decoder_hidden_states (`tuple(torch.FloatTensor)`, *optional*, returned when `output_hidden_states=True` is passed or when `config.output_hidden_states=True`): + Tuple of `torch.FloatTensor` (one for the output of the embeddings + one for the output of each layer) of + shape `(batch_size, decoder_sequence_length, hidden_size)`. + + Hidden-states of main stream of the decoder at the output of each layer plus the initial embedding outputs. + decoder_ngram_hidden_states (`tuple(torch.FloatTensor)`, *optional*, returned when `output_hidden_states=True` is passed or when `config.output_hidden_states=True`): + Tuple of `torch.FloatTensor` (one for the output of the embeddings + one for the output of each layer) of + shape `(batch_size, ngram * decoder_sequence_length, hidden_size)`. + + Hidden-states of the predict stream of the decoder at the output of each layer plus the initial embedding + outputs. + decoder_attentions (`tuple(torch.FloatTensor)`, *optional*, returned when `output_attentions=True` is passed or when `config.output_attentions=True`): + Tuple of `torch.FloatTensor` (one for each layer) of shape `(batch_size, num_attn_heads, + decoder_sequence_length, decoder_sequence_length)`. + + Attentions weights of the decoder, after the attention softmax, used to compute the weighted average in the + self-attention heads. + decoder_ngram_attentions (`tuple(torch.FloatTensor)`, *optional*, returned when `output_attentions=True` is passed or when `config.output_attentions=True`): + Tuple of `torch.FloatTensor` (one for each layer) of shape `(batch_size, num_attn_heads, + decoder_sequence_length, decoder_sequence_length)`. + + Attentions weights of the predict stream of the decoder, after the attention softmax, used to compute the + weighted average in the self-attention heads. + cross_attentions (`tuple(torch.FloatTensor)`, *optional*, returned when `output_attentions=True` is passed or when `config.output_attentions=True`): + Tuple of `torch.FloatTensor` (one for each layer) of shape `(batch_size, num_attn_heads, + encoder_sequence_length, decoder_sequence_length)`. + + Attentions weights of the cross-attention layer of the decoder, after the attention softmax, used to + compute the weighted average in the + encoder_last_hidden_state (`torch.FloatTensor` of shape `(batch_size, encoder_sequence_length, hidden_size)`, *optional*): + Sequence of hidden-states at the output of the last layer of the encoder of the model. + encoder_hidden_states (`tuple(torch.FloatTensor)`, *optional*, returned when `output_hidden_states=True` is passed or when `config.output_hidden_states=True`): + Tuple of `torch.FloatTensor` (one for the output of the embeddings + one for the output of each layer) of + shape `(batch_size, encoder_sequence_length, hidden_size)`. + + Hidden-states of the encoder at the output of each layer plus the initial embedding outputs. + encoder_attentions (`tuple(torch.FloatTensor)`, *optional*, returned when `output_attentions=True` is passed or when `config.output_attentions=True`): + Tuple of `torch.FloatTensor` (one for each layer) of shape `(batch_size, num_attn_heads, + encoder_sequence_length, encoder_sequence_length)`. Attentions weights of the encoder, after the attention + softmax, used to compute the weighted average in the self-attention heads. + """ + + loss: Optional[torch.FloatTensor] = None + logits: torch.FloatTensor = None + logits_ngram: Optional[torch.FloatTensor] = None + past_key_values: Optional[Tuple[torch.FloatTensor]] = None + decoder_hidden_states: Optional[Tuple[torch.FloatTensor]] = None + decoder_ngram_hidden_states: Optional[Tuple[torch.FloatTensor]] = None + decoder_attentions: Optional[Tuple[torch.FloatTensor]] = None + decoder_ngram_attentions: Optional[Tuple[torch.FloatTensor]] = None + cross_attentions: Optional[Tuple[torch.FloatTensor]] = None + encoder_last_hidden_state: Optional[torch.FloatTensor] = None + encoder_hidden_states: Optional[Tuple[torch.FloatTensor]] = None + encoder_attentions: Optional[Tuple[torch.FloatTensor]] = None + + @property + def decoder_cross_attentions(self): + warnings.warn( + "`decoder_cross_attentions` is deprecated and will be removed soon. Please use `cross_attentions`" + " instead.", + FutureWarning, + ) + return self.cross_attentions + + +@dataclass +class ProphetNetSeq2SeqModelOutput(ModelOutput): + """ + Base class for model encoder's outputs that also contains : pre-computed hidden states that can speed up sequential + decoding. + + Args: + last_hidden_state (`torch.FloatTensor` of shape `(batch_size, decoder_sequence_length, hidden_size)`): + Sequence of main stream hidden-states at the output of the last layer of the decoder of the model. + + If `past_key_values` is used only the last hidden-state of the sequences of shape `(batch_size, 1, + hidden_size)` is output. + last_hidden_state_ngram (`torch.FloatTensor` of shape `(batch_size,ngram * decoder_sequence_length, config.vocab_size)`, *optional*): + Sequence of predict stream hidden-states at the output of the last layer of the decoder of the model. + past_key_values (`List[torch.FloatTensor]`, *optional*, returned when `use_cache=True` is passed or when `config.use_cache=True`): + List of `torch.FloatTensor` of length `config.n_layers`, with each tensor of shape `(2, batch_size, + num_attn_heads, decoder_sequence_length, embed_size_per_head)`). + + Contains pre-computed hidden-states (key and values in the attention blocks) of the decoder that can be + used (see `past_key_values` input) to speed up sequential decoding. + decoder_hidden_states (`tuple(torch.FloatTensor)`, *optional*, returned when `output_hidden_states=True` is passed or when `config.output_hidden_states=True`): + Tuple of `torch.FloatTensor` (one for the output of the embeddings + one for the output of each layer) of + shape `(batch_size, decoder_sequence_length, hidden_size)`. + + Hidden-states of main stream of the decoder at the output of each layer plus the initial embedding outputs. + decoder_ngram_hidden_states (`tuple(torch.FloatTensor)`, *optional*, returned when `output_hidden_states=True` is passed or when `config.output_hidden_states=True`): + Tuple of `torch.FloatTensor` (one for the output of the embeddings + one for the output of each layer) of + shape `(batch_size, ngram * decoder_sequence_length, hidden_size)`. + + Hidden-states of the predict stream of the decoder at the output of each layer plus the initial embedding + outputs. + decoder_attentions (`tuple(torch.FloatTensor)`, *optional*, returned when `output_attentions=True` is passed or when `config.output_attentions=True`): + Tuple of `torch.FloatTensor` (one for each layer) of shape `(batch_size, num_attn_heads, + decoder_sequence_length, decoder_sequence_length)`. + + Attentions weights of the decoder, after the attention softmax, used to compute the weighted average in the + self-attention heads. + decoder_ngram_attentions (`tuple(torch.FloatTensor)`, *optional*, returned when `output_attentions=True` is passed or when `config.output_attentions=True`): + Tuple of `torch.FloatTensor` (one for each layer) of shape `(batch_size, num_attn_heads, + decoder_sequence_length, decoder_sequence_length)`. + + Attentions weights of the predict stream of the decoder, after the attention softmax, used to compute the + weighted average in the + cross_attentions (`tuple(torch.FloatTensor)`, *optional*, returned when `output_attentions=True` is passed or when `config.output_attentions=True`): + Tuple of `torch.FloatTensor` (one for each layer) of shape `(batch_size, num_attn_heads, + encoder_sequence_length, decoder_sequence_length)`. + + Attentions weights of the cross-attention layer of the decoder, after the attention softmax, used to + compute the weighted average in the + encoder_last_hidden_state (`torch.FloatTensor` of shape `(batch_size, encoder_sequence_length, hidden_size)`, *optional*): + Sequence of hidden-states at the output of the last layer of the encoder of the model. + encoder_hidden_states (`tuple(torch.FloatTensor)`, *optional*, returned when `output_hidden_states=True` is passed or when `config.output_hidden_states=True`): + Tuple of `torch.FloatTensor` (one for the output of the embeddings + one for the output of each layer) of + shape `(batch_size, encoder_sequence_length, hidden_size)`. + + Hidden-states of the encoder at the output of each layer plus the initial embedding outputs. + encoder_attentions (`tuple(torch.FloatTensor)`, *optional*, returned when `output_attentions=True` is passed or when `config.output_attentions=True`): + Tuple of `torch.FloatTensor` (one for each layer) of shape `(batch_size, num_attn_heads, + encoder_sequence_length, encoder_sequence_length)`. + + Attentions weights of the encoder, after the attention softmax, used to compute the weighted average in the + self-attention heads. + """ + + last_hidden_state: torch.FloatTensor + last_hidden_state_ngram: Optional[torch.FloatTensor] = None + past_key_values: Optional[Tuple[torch.FloatTensor]] = None + decoder_hidden_states: Optional[Tuple[torch.FloatTensor]] = None + decoder_ngram_hidden_states: Optional[Tuple[torch.FloatTensor]] = None + decoder_attentions: Optional[Tuple[torch.FloatTensor]] = None + decoder_ngram_attentions: Optional[Tuple[torch.FloatTensor]] = None + cross_attentions: Optional[Tuple[torch.FloatTensor]] = None + encoder_last_hidden_state: Optional[torch.FloatTensor] = None + encoder_hidden_states: Optional[Tuple[torch.FloatTensor]] = None + encoder_attentions: Optional[Tuple[torch.FloatTensor]] = None + + @property + def decoder_cross_attentions(self): + warnings.warn( + "`decoder_cross_attentions` is deprecated and will be removed soon. Please use `cross_attentions`" + " instead.", + FutureWarning, + ) + return self.cross_attentions + + +@dataclass +class ProphetNetDecoderModelOutput(ModelOutput): + """ + Base class for model's outputs that may also contain a past key/values (to speed up sequential decoding). + + Args: + last_hidden_state (`torch.FloatTensor` of shape `(batch_size, decoder_sequence_length, hidden_size)`): + Sequence of main stream hidden-states at the output of the last layer of the decoder of the model. + + If `past_key_values` is used only the last hidden-state of the sequences of shape `(batch_size, 1, + hidden_size)` is output. + last_hidden_state_ngram (`torch.FloatTensor` of shape `(batch_size, ngram * decoder_sequence_length, config.vocab_size)`): + Sequence of predict stream hidden-states at the output of the last layer of the decoder of the model. + past_key_values (`List[torch.FloatTensor]`, *optional*, returned when `use_cache=True` is passed or when `config.use_cache=True`): + List of `torch.FloatTensor` of length `config.n_layers`, with each tensor of shape `(2, batch_size, + num_attn_heads, decoder_sequence_length, embed_size_per_head)`). + + Contains pre-computed hidden-states (key and values in the attention blocks) of the decoder that can be + used (see `past_key_values` input) to speed up sequential decoding. + hidden_states (`tuple(torch.FloatTensor)`, *optional*, returned when `output_hidden_states=True` is passed or when `config.output_hidden_states=True`): + Tuple of `torch.FloatTensor` (one for the output of the embeddings + one for the output of each layer) of + shape `(batch_size, decoder_sequence_length, hidden_size)`. + + Hidden-states of main stream of the decoder at the output of each layer plus the initial embedding outputs. + ngram_hidden_states (`tuple(torch.FloatTensor)`, *optional*, returned when `output_hidden_states=True` is passed or when `config.output_hidden_states=True`): + Tuple of `torch.FloatTensor` (one for the output of the embeddings + one for the output of each layer) of + shape `(batch_size, ngram * decoder_sequence_length, hidden_size)`. + + Hidden-states of the predict stream of the decoder at the output of each layer plus the initial embedding + outputs. + attentions (`tuple(torch.FloatTensor)`, *optional*, returned when `output_attentions=True` is passed or when `config.output_attentions=True`): + Tuple of `torch.FloatTensor` (one for each layer) of shape `(batch_size, num_attn_heads, + decoder_sequence_length, decoder_sequence_length)`. + + Attentions weights of the decoder, after the attention softmax, used to compute the weighted average in the + self-attention heads. + ngram_attentions (`tuple(torch.FloatTensor)`, *optional*, returned when `output_attentions=True` is passed or when `config.output_attentions=True`): + Tuple of `torch.FloatTensor` (one for each layer) of shape `(batch_size, num_attn_heads, + decoder_sequence_length, decoder_sequence_length)`. + + Attentions weights of the predict stream of the decoder, after the attention softmax, used to compute the + weighted average in the + cross_attentions (`tuple(torch.FloatTensor)`, *optional*, returned when `output_attentions=True` is passed or when `config.output_attentions=True`): + Tuple of `torch.FloatTensor` (one for each layer) of shape `(batch_size, num_attn_heads, + encoder_sequence_length, decoder_sequence_length)`. + + Attentions weights of the cross-attention layer of the decoder, after the attention softmax, used to + compute the weighted average in the + """ + + last_hidden_state: torch.FloatTensor + last_hidden_state_ngram: Optional[torch.FloatTensor] = None + past_key_values: Optional[Tuple[torch.FloatTensor]] = None + hidden_states: Optional[Tuple[torch.FloatTensor]] = None + hidden_states_ngram: Optional[Tuple[torch.FloatTensor]] = None + attentions: Optional[Tuple[torch.FloatTensor]] = None + ngram_attentions: Optional[Tuple[torch.FloatTensor]] = None + cross_attentions: Optional[Tuple[torch.FloatTensor]] = None + + +@dataclass +class ProphetNetDecoderLMOutput(ModelOutput): + """ + Base class for model's outputs that may also contain a past key/values (to speed up sequential decoding). + + Args: + loss (`torch.FloatTensor` of shape `(1,)`, *optional*, returned when `labels` is provided): + Language modeling loss. + logits (`torch.FloatTensor` of shape `(batch_size, decoder_sequence_length, config.vocab_size)`): + Prediction scores of the main stream language modeling head (scores for each vocabulary token before + SoftMax). + logits_ngram (`torch.FloatTensor` of shape `(batch_size, ngram * decoder_sequence_length, config.vocab_size)`): + Prediction scores of the predict stream language modeling head (scores for each vocabulary token before + SoftMax). + past_key_values (`List[torch.FloatTensor]`, *optional*, returned when `use_cache=True` is passed or when `config.use_cache=True`): + List of `torch.FloatTensor` of length `config.n_layers`, with each tensor of shape `(2, batch_size, + num_attn_heads, decoder_sequence_length, embed_size_per_head)`). + + Contains pre-computed hidden-states (key and values in the attention blocks) of the decoder that can be + used (see `past_key_values` input) to speed up sequential decoding. + hidden_states (`tuple(torch.FloatTensor)`, *optional*, returned when `output_hidden_states=True` is passed or when `config.output_hidden_states=True`): + Tuple of `torch.FloatTensor` (one for the output of the embeddings + one for the output of each layer) of + shape `(batch_size, decoder_sequence_length, hidden_size)`. + + Hidden-states of main stream of the decoder at the output of each layer plus the initial embedding outputs. + ngram_hidden_states (`tuple(torch.FloatTensor)`, *optional*, returned when `output_hidden_states=True` is passed or when `config.output_hidden_states=True`): + Tuple of `torch.FloatTensor` (one for the output of the embeddings + one for the output of each layer) of + shape `(batch_size, ngram * decoder_sequence_length, hidden_size)`. + + Hidden-states of the predict stream of the decoder at the output of each layer plus the initial embedding + outputs. + attentions (`tuple(torch.FloatTensor)`, *optional*, returned when `output_attentions=True` is passed or when `config.output_attentions=True`): + Tuple of `torch.FloatTensor` (one for each layer) of shape `(batch_size, num_attn_heads, + decoder_sequence_length, decoder_sequence_length)`. + + Attentions weights of the decoder, after the attention softmax, used to compute the weighted average in the + self-attention heads. + ngram_attentions (`tuple(torch.FloatTensor)`, *optional*, returned when `output_attentions=True` is passed or when `config.output_attentions=True`): + Tuple of `torch.FloatTensor` (one for each layer) of shape `(batch_size, num_attn_heads, + decoder_sequence_length, decoder_sequence_length)`. + + Attentions weights of the predict stream of the decoder, after the attention softmax, used to compute the + weighted average in the + cross_attentions (`tuple(torch.FloatTensor)`, *optional*, returned when `output_attentions=True` is passed or when `config.output_attentions=True`): + Tuple of `torch.FloatTensor` (one for each layer) of shape `(batch_size, num_attn_heads, + encoder_sequence_length, decoder_sequence_length)`. + + Attentions weights of the cross-attention layer of the decoder, after the attention softmax, used to + compute the weighted average in the + """ + + loss: Optional[torch.FloatTensor] = None + logits: torch.FloatTensor = None + logits_ngram: Optional[torch.FloatTensor] = None + past_key_values: Optional[Tuple[torch.FloatTensor]] = None + hidden_states: Optional[Tuple[torch.FloatTensor]] = None + hidden_states_ngram: Optional[Tuple[torch.FloatTensor]] = None + attentions: Optional[Tuple[torch.FloatTensor]] = None + ngram_attentions: Optional[Tuple[torch.FloatTensor]] = None + cross_attentions: Optional[Tuple[torch.FloatTensor]] = None + + +class ProphetNetPreTrainedModel(PreTrainedModel): + config_class = ProphetNetConfig + base_model_prefix = "prophetnet" + supports_gradient_checkpointing = True + + def _init_weights(self, module): + if isinstance(module, nn.Linear): + module.weight.data.normal_(mean=0.0, std=self.config.init_std) + if module.bias is not None: + module.bias.data.zero_() + elif isinstance(module, nn.Embedding): + module.weight.data.normal_(mean=0.0, std=self.config.init_std) + if module.padding_idx is not None: + module.weight.data[module.padding_idx].zero_() + + def _shift_right(self, input_ids): + decoder_start_token_id = self.config.decoder_start_token_id + pad_token_id = self.config.pad_token_id + + assert decoder_start_token_id is not None, ( + "self.model.config.decoder_start_token_id has to be defined. In ProphetNet it is usually set to the" + " pad_token_id. See ProphetNet docs for more information" + ) + + # shift inputs to the right + shifted_input_ids = input_ids.new_zeros(input_ids.shape) + shifted_input_ids[..., 1:] = input_ids[..., :-1].clone() + shifted_input_ids[..., 0] = decoder_start_token_id + + assert pad_token_id is not None, "self.model.config.pad_token_id has to be defined." + # replace possible -100 values in labels by `pad_token_id` + shifted_input_ids.masked_fill_(shifted_input_ids == -100, pad_token_id) + + assert torch.all(shifted_input_ids >= 0).item(), "Verify that `shifted_input_ids` has only positive values" + + return shifted_input_ids + + +class ProphetNetPositionalEmbeddings(nn.Embedding): + """ + This module learns positional embeddings up to a fixed maximum size. Padding ids are ignored by either offsetting + based on padding_idx or by setting padding_idx to None and ensuring that the appropriate position ids are passed to + the forward function. + """ + + def __init__(self, config: ProphetNetConfig) -> None: + self.max_length = config.max_position_embeddings + super().__init__(config.max_position_embeddings, config.hidden_size, config.pad_token_id) + + def forward(self, inputs_shape, device, attention_mask=None, past_key_values=None, position_ids=None): + assert (position_ids is None) or ( + self.padding_idx is None + ), "If position_ids is pre-computed then padding_idx should not be set." + + if position_ids is None: + if past_key_values is not None: + # position_ids is the same for every token when decoding a single step + # Without the int() cast, it doesn't work in some cases when exporting to ONNX + prev_num_input_ids = past_key_values[0][0].shape[2] + num_input_ids = inputs_shape[1] + prev_num_input_ids + position_ids = torch.ones((1, 1), dtype=torch.long, device=device) * ( + int(self.padding_idx + num_input_ids) + ) + else: + if attention_mask is None: + attention_mask = torch.ones(inputs_shape, dtype=torch.long, device=device) + + # retrieve position_ids from input_ids / attention_mask + position_ids = ( + torch.cumsum(attention_mask, dim=1).type_as(attention_mask) * attention_mask + ).long() + self.padding_idx + + # make sure position_ids are not bigger then max_length + position_ids = position_ids.clamp(0, self.max_length - 1) + + return super().forward(position_ids), position_ids + + def _forward(self, position_ids): + return super().forward(position_ids) + + +class ProphetNetAttention(nn.Module): + """Multi-headed attention from 'Attention Is All You Need' paper""" + + def __init__( + self, + config: ProphetNetConfig, + num_attn_heads: int, + ): + super().__init__() + hidden_size = config.hidden_size + + self.attention_dropout = config.attention_dropout + self.dropout = config.dropout + self.num_attn_heads = num_attn_heads + self.head_dim = hidden_size // num_attn_heads + + assert self.head_dim * num_attn_heads == hidden_size, ( + "`config.hidden_size` must be divisible by `config.num_encoder_attention_heads` and" + " `config.num_decoder_attention_heads`" + ) + + self.key_proj = nn.Linear(hidden_size, hidden_size) + self.value_proj = nn.Linear(hidden_size, hidden_size) + self.query_proj = nn.Linear(hidden_size, hidden_size) + + self.out_proj = nn.Linear(hidden_size, hidden_size) + + def _shape(self, tensor: torch.Tensor, seq_len: int, bsz: int): + return tensor.view(bsz, seq_len, self.num_attn_heads, self.head_dim).transpose(1, 2).contiguous() + + def forward( + self, + hidden_states, + key_value_states: Optional[Tensor] = None, + attention_mask: Optional[Tensor] = None, + layer_head_mask: Optional[Tensor] = None, + past_key_value: Optional[Tuple[Tensor]] = None, + output_attentions: bool = False, + ) -> Tuple[Tensor, Optional[Tensor]]: + batch_size, tgt_len, hidden_size = hidden_states.size() + + # if key_value_states are provided this layer is used as a cross-attention layer + # for the decoder + is_cross_attention = key_value_states is not None + assert list(hidden_states.size()) == [ + batch_size, + tgt_len, + hidden_size, + ], f"Size of hidden states should be {batch_size, tgt_len, hidden_size}, but is {hidden_states.size()}" + + # previous time steps are cached - no need to recompute key and value if they are static + query_states = self.query_proj(hidden_states) / (self.head_dim**0.5) + + if is_cross_attention and past_key_value is not None: + # reuse k,v, cross_attentions + key_states = past_key_value[0] + value_states = past_key_value[1] + elif is_cross_attention: + # cross_attentions + key_states = self._shape(self.key_proj(key_value_states), -1, batch_size) + value_states = self._shape(self.value_proj(key_value_states), -1, batch_size) + else: + # self_attention + key_states = self._shape(self.key_proj(hidden_states), -1, batch_size) + value_states = self._shape(self.value_proj(hidden_states), -1, batch_size) + + if is_cross_attention: + # if cross_attention save Tuple(torch.Tensor, torch.Tensor) of all cross attention key/value_states. + # Further calls to cross_attention layer can then reuse all cross-attention + # key/value_states (first "if" case) + # if encoder bi-directional self-attention `past_key_value` is always `None` + past_key_value = (key_states, value_states) + + # project states into the correct shape + proj_shape = (batch_size, self.num_attn_heads, -1, self.head_dim) + query_states = self._shape(query_states, tgt_len, batch_size).view(*proj_shape) + key_states = key_states.view(*proj_shape) + value_states = value_states.view(*proj_shape) + src_len = key_states.size(2) + attn_weights = torch.einsum("bsij,bsjk->bsik", query_states, key_states.transpose(2, 3)) + expected_shape = (batch_size, self.num_attn_heads, tgt_len, src_len) + if attn_weights.size() != expected_shape: + raise ValueError(f"Attention weights should have size {expected_shape}, but is {attn_weights.size()}") + + # This is part of a workaround to get around fork/join parallelism not supporting Optional types. + if attention_mask is not None and attention_mask.dim() == 0: + attention_mask = None + + expected_shape = (batch_size, self.num_attn_heads, 1, src_len) + if attention_mask is not None and attention_mask.size() != expected_shape: + raise ValueError(f"Attention mask should have size {expected_shape}, but is {attention_mask.size()}") + if attention_mask is not None: # don't attend to padding symbols + attn_weights = attn_weights + attention_mask + if output_attentions: + attn_weights_reshaped = attn_weights + else: + attn_weights_reshaped = None + + attn_weights = nn.functional.softmax(attn_weights, dim=-1) + + if layer_head_mask is not None: + assert layer_head_mask.size() == (self.num_attn_heads,), ( + f"Head mask for a single layer should be of size {(self.num_attn_heads,)}, but is" + f" {layer_head_mask.size()}" + ) + attn_weights = layer_head_mask.view(1, -1, 1, 1) * attn_weights.view( + batch_size, self.num_attn_heads, tgt_len, src_len + ) + + # apply head_mask also on attn_weights_reshaped which is used for n-gram attention inside the model + attn_weights_reshaped = layer_head_mask.view(1, -1, 1, 1) * attn_weights_reshaped + + attn_probs = nn.functional.dropout( + attn_weights, + p=self.attention_dropout, + training=self.training, + ) + attn_output = torch.einsum("bsij,bsjk->bsik", attn_probs, value_states) + expected_shape = (batch_size, self.num_attn_heads, tgt_len, self.head_dim) + if attn_output.size() != expected_shape: + raise ValueError(f"`attn_output` should have shape {expected_shape}, but is of shape {attn_output.size()}") + + attn_output = attn_output.transpose(1, 2).reshape(batch_size, tgt_len, hidden_size) + attn_output = self.out_proj(attn_output) + + attn_output = nn.functional.dropout(attn_output, p=self.dropout, training=self.training) + return attn_output, attn_weights_reshaped, past_key_value + + +class ProphetNetFeedForward(nn.Module): + """ + This is the residual two feed-forward layer block based on the original Transformer implementation. + """ + + def __init__(self, config: ProphetNetConfig, ffn_dim: int): + super().__init__() + self.activation_fn = ACT2FN[config.activation_function] + self.intermediate = nn.Linear(config.hidden_size, ffn_dim) + self.output = nn.Linear(ffn_dim, config.hidden_size) + self.activation_dropout = config.activation_dropout + self.dropout = config.dropout + + def forward(self, hidden_states): + hidden_states = self.intermediate(hidden_states) + hidden_states = self.activation_fn(hidden_states) + + hidden_states = nn.functional.dropout(hidden_states, p=self.activation_dropout, training=self.training) + hidden_states = self.output(hidden_states) + hidden_states = nn.functional.dropout(hidden_states, p=self.dropout, training=self.training) + return hidden_states + + +class ProphetNetNgramSelfAttention(nn.Module): + def __init__(self, config: ProphetNetConfig): + super().__init__() + self.hidden_size = config.hidden_size + + self.num_buckets = config.num_buckets + self.relative_max_distance = config.relative_max_distance + self.num_attn_heads = config.num_decoder_attention_heads + self.dropout = config.dropout + self.attention_dropout = config.attention_dropout + self.head_dim = config.hidden_size // self.num_attn_heads + self.ngram = config.ngram + + assert ( + self.head_dim * self.num_attn_heads == config.hidden_size + ), "config.hidden_size must be divisible by num_attn_heads" + # key, value, query projection + self.key_proj = nn.Linear(config.hidden_size, config.hidden_size) + self.value_proj = nn.Linear(config.hidden_size, config.hidden_size) + self.query_proj = nn.Linear(config.hidden_size, config.hidden_size) + + # out projection + self.out_proj = nn.Linear(config.hidden_size, config.hidden_size) + + # rel position embeddings + self.relative_pos_embeddings = nn.Linear(config.hidden_size, self.num_buckets * self.num_attn_heads) + + # for onnx runtime + self.onnx_trace = False + + def _shape(self, tensor, seq_len, batch_size): + return tensor.view(batch_size, seq_len, self.num_attn_heads, self.head_dim).transpose(1, 2).contiguous() + + def prepare_for_onnx_export_(self): + self.onnx_trace = True + + def forward( + self, + hidden_states, + past_key_value: Optional[Tuple[Tensor]] = None, + attention_mask=None, + layer_head_mask=None, + extended_predict_attention_mask=None, + main_relative_position_buckets=None, + predict_relative_position_buckets=None, + position_ids=None, + ): + batch_size, ngram_sequence_length, hidden_size = hidden_states.size() + assert list(hidden_states.size()) == [batch_size, ngram_sequence_length, hidden_size], ( + f"`hidden_states` should be of shape {batch_size, ngram_sequence_length, hidden_size}, but is of shape" + f" {hidden_states.shape}" + ) + + # project + query_states = self.query_proj(hidden_states) + key_states = self.key_proj(hidden_states) + value_states = self.value_proj(hidden_states) + + # normalize + query_states = query_states / (self.head_dim**0.5) + + # reshape + query_states = self._shape(query_states, ngram_sequence_length, batch_size) + key_states = self._shape(key_states, -1, batch_size) + value_states = self._shape(value_states, -1, batch_size) + proj_shape = (batch_size, self.num_attn_heads, -1, self.head_dim) + + query_states = query_states.view(*proj_shape) + key_states = key_states.view(*proj_shape) + value_states = value_states.view(*proj_shape) + + # chunk into main stream and predict stream + hidden_states_list = hidden_states.chunk(1 + self.ngram, dim=1) + query_states_list = query_states.chunk(1 + self.ngram, dim=2) + key_states_list = key_states.chunk(1 + self.ngram, dim=2) + value_states_list = value_states.chunk(1 + self.ngram, dim=2) + + main_hidden_states, hidden_states_predict_list = hidden_states_list[0], hidden_states_list[1:] + main_query_states, predict_query_states_list = query_states_list[0], query_states_list[1:] + main_key_states, predict_key_states_list = key_states_list[0], key_states_list[1:] + main_value_states, predict_value_states_list = value_states_list[0], value_states_list[1:] + + # saved states are stored with shape (batch_size, num_attn_heads, seq_len, head_dim) + if past_key_value is not None: + prev_main_key_states = past_key_value[0] + main_key_states = torch.cat((prev_main_key_states, main_key_states), dim=2) + prev_main_value_states = past_key_value[1] + main_value_states = torch.cat((prev_main_value_states, main_value_states), dim=2) + + # Update cache + past_key_value = (main_key_states, main_value_states) + + # get seq_length of main stream only + sequence_length = ngram_sequence_length // (1 + self.ngram) + + # MAIN-STREAM + # main attn weights + # [batch_size, number_heads, sequence_length, head_dimesion] + # x [batch_size, number_heads, head_dimesion, sequence_length] + # -> [batch_size, number_heads, sequence_length, sequence_length] + main_attn_weights = torch.einsum("bntc,bncs->bnts", main_query_states, main_key_states.transpose(2, 3)) + + # retrieve relative position embeddings for each layer -> see paper for more details + main_relative_pos_embeddings = self.get_main_relative_pos_embeddings( + main_hidden_states, main_attn_weights, position_ids, main_relative_position_buckets + ) + + main_attn_weights = main_attn_weights + main_relative_pos_embeddings + + if attention_mask is not None: + main_attn_weights = main_attn_weights + attention_mask + + main_attn_probs = softmax( + main_attn_weights, + dim=-1, + onnx_trace=self.onnx_trace, + ).type_as(main_attn_weights) + + if layer_head_mask is not None: + assert layer_head_mask.size() == (self.num_attn_heads,), ( + f"Head mask for a single layer should be of size {(self.num_attn_heads,)}, but is" + f" {layer_head_mask.size()}" + ) + main_attn_probs = layer_head_mask.view(1, -1, 1, 1) * main_attn_probs.view( + batch_size, self.num_attn_heads, -1, sequence_length + ) + + main_attn_probs = nn.functional.dropout(main_attn_probs, p=self.attention_dropout, training=self.training) + # project to attn_output + # [batch_size, number_heads, sequence_length, sequence_length] + # x [batch_size, number_heads, sequence_length, head_dimesion] + # -> [batch_size, number_heads, sequence_length, head_dimesion] + main_attn_output = torch.einsum("bntc,bncs->bnts", main_attn_probs, main_value_states) + # reshape so that num_heads dim is merged into last `head_dim` axis + main_attn_output = main_attn_output.transpose(1, 2).reshape(batch_size, 1, sequence_length, hidden_size) + main_attn_output = self.out_proj(main_attn_output) + + # PREDICT-STREAM + # [batch_size, ngram, number_heads, sequence_length, head_dimesion] + predict_query_states = torch.stack(predict_query_states_list, 1).view( + batch_size, self.ngram, self.num_attn_heads, sequence_length, self.head_dim + ) + + # [batch_size, ngram, number_heads, 2*sequence_length, head_dimesion] + predict_key_states = torch.stack([torch.cat([main_key_states, key], 2) for key in predict_key_states_list], 1) + + # [batch_size, sequence_length, ngram, hidden_size] + predict_hidden_states = torch.stack(hidden_states_predict_list, dim=2) + + # [batch_size, number_heads, ngram, 2*sequence_length, head_dimesion] + predict_value_states = torch.cat( + [torch.cat([main_value_states, v_p], 2).unsqueeze(2) for v_p in predict_value_states_list], 2 + ) + + # [batch_size, ngram, number_heads, sequence_length, head_dimesion] + # x [batch_size, ngram, number_heads, 2*sequence_length, head_dimesion] + # -> [batch_size, ngram, number_heads, sequence_length, 2*sequence_length] + predict_attn_weights = torch.einsum("bnhtc,bnhsc->bnhts", (predict_query_states, predict_key_states)) + + # retrieve relative position embeddings for each layer -> see paper for more details + # [batch_size, ngram, number_heads, sequence_length, predict_relative_pos_embeddings] + predict_relative_pos_embeddings = self.get_predict_relative_pos_embeddings( + predict_hidden_states, predict_attn_weights, position_ids, predict_relative_position_buckets + ) + + # [batch_size, ngram, number_heads, sequence_length, 2*sequence_length] + predict_attn_weights = predict_attn_weights + predict_relative_pos_embeddings + + if extended_predict_attention_mask is not None: + # Permuting Predict attention mask to [batch_size, ngram, number_heads, sequence_length, 2*sequence_length] + extended_predict_attention_mask = extended_predict_attention_mask.permute(0, 2, 1, 3, 4) + extended_predict_attention_mask = extended_predict_attention_mask.to(predict_attn_weights.dtype) + predict_attn_weights = predict_attn_weights + extended_predict_attention_mask + + predict_attn_probs = softmax( + predict_attn_weights, + dim=-1, + onnx_trace=self.onnx_trace, + ).type_as(predict_attn_weights) + + if layer_head_mask is not None: + assert layer_head_mask.size() == (self.num_attn_heads,), ( + f"Head mask for a single layer should be of size {(self.num_attn_heads,)}, but is" + f" {layer_head_mask.size()}" + ) + predict_attn_probs = layer_head_mask.view(1, 1, -1, 1, 1) * predict_attn_probs + + predict_attn_probs = nn.functional.dropout( + predict_attn_probs, p=self.attention_dropout, training=self.training + ) + # project to attention output + # [batch_size, ngram, number_heads, sequence_length, 2*sequence_length] + # x [batch_size, ngram, number_heads, 2*sequence_length, head_dimesion] + # -> [batch_size, ngram, number_heads, sequence_length, head_dimesion] + predict_attn_output = torch.einsum( + "bnhts,bnhsc->bnhtc", (predict_attn_probs, predict_value_states.transpose(1, 2)) + ) + + # reshape so that num_heads dim is merged into last `head_dim` axis + # [batch_size, ngram, number_heads, sequence_length, head_dimesion] -> [batch_size, ngram, sequence_length, hidden_size] + predict_attn_output = predict_attn_output.transpose(2, 3) + predict_attn_output = predict_attn_output.reshape(batch_size, self.ngram, sequence_length, hidden_size) + predict_attn_output = self.out_proj(predict_attn_output) + + # concat to single attn output + # [batch_size, (1+ngram)*sequence_length, hidden_size] + attn_output = torch.cat([main_attn_output, predict_attn_output], 1).view(batch_size, -1, hidden_size) + # reshape into better form for `config.output_attentions` + main_attn_probs = main_attn_probs.view(batch_size, self.num_attn_heads, sequence_length, -1) + + attn_output = nn.functional.dropout(attn_output, p=self.dropout, training=self.training) + + return attn_output, main_attn_probs, predict_attn_probs, past_key_value + + def get_main_relative_pos_embeddings( + self, hidden_states, attn_weights, position_ids, main_relative_position_buckets + ): + # input hidden_states [batch_size, sequence_length, hidden_size] + # input attn_weights [batch_size, num_heads, sequence_length, sequence_length] + # input position_ids [batch_size, sequence_length] or [1,1] + batch_size, num_attn_heads, tgt_len, src_len = attn_weights.shape + attn_weights = attn_weights.view(batch_size, num_attn_heads, tgt_len, src_len) + if main_relative_position_buckets is None: + batch_size, sequence_length = hidden_states.shape[:2] + relative_positions = ( + torch.arange(1, attn_weights.shape[-1] + 1) + .unsqueeze(0) + .unsqueeze(0) + .repeat(batch_size, sequence_length, 1) + .to(position_ids.device) + ) + # [batch_size, sequence_length, sequence_length+1] + relative_positions = relative_positions - position_ids.unsqueeze(0).repeat(batch_size, sequence_length, 1) + main_relative_position_buckets = compute_relative_buckets( + self.num_buckets, self.relative_max_distance, relative_positions, False + ) + + # [batch_size, sequence_length, num_buckets * num_heads] + rel_pos_embeddings = self.relative_pos_embeddings(hidden_states) + rel_pos_embeddings = rel_pos_embeddings.view( + rel_pos_embeddings.shape[:2] + (self.num_buckets, self.num_attn_heads) + ) + rel_pos_embeddings = rel_pos_embeddings.permute(0, 3, 1, 2) + # [batch_size, num_heads, sequence_length, num_buckets] + rel_pos_embeddings = rel_pos_embeddings.reshape(attn_weights.shape[:3] + (-1,)) + + main_relative_position_buckets = main_relative_position_buckets.repeat(1, self.num_attn_heads, 1) + # [batch_size * num_heads * sequence_length, sequence_length] + main_relative_position_buckets = main_relative_position_buckets.view( + -1, main_relative_position_buckets.shape[-1] + ) + main_relative_position_buckets = main_relative_position_buckets.long() + # [batch_size * num_heads * sequence_length, sequence_length] + rel_pos_embeddings = rel_pos_embeddings.reshape(-1, rel_pos_embeddings.size(-1)) + + main_relative_pos_embeddings = torch.gather(rel_pos_embeddings, dim=1, index=main_relative_position_buckets) + main_relative_pos_embeddings = main_relative_pos_embeddings.view(batch_size, num_attn_heads, tgt_len, -1) + return main_relative_pos_embeddings + + def get_predict_relative_pos_embeddings( + self, hidden_states, attn_weights, position_ids, predict_relative_position_buckets + ): + # input hidden_states [batch_size, sequence_length, ngram, hidden_size] + # input attn_weights [batch_size, ngram, num_heads, sequence_length, 2*sequence_length] + # input position_ids [batch_size, sequence_length] or [1,1] + # input predict_relative_position_buckets [batch_size, sequence_length, 2*sequence_length] or None + batch_size, sequence_length = hidden_states.shape[0:2] + + if predict_relative_position_buckets is None: + key_sequence_length = attn_weights.shape[-1] + assert ( + position_ids[0][0] == key_sequence_length - 1 + ), "`position_ids` are incorrect. They should be of the format 1 2 3 4 5 ... (key_sequence_length - 1)" + relative_positions = ( + torch.arange(0, key_sequence_length) + .unsqueeze(0) + .unsqueeze(0) + .repeat(batch_size, sequence_length, 1) + .to(position_ids.device) + ) + + relative_positions = relative_positions - position_ids.unsqueeze(0).repeat(batch_size, sequence_length, 1) + predict_relative_position_buckets = compute_relative_buckets( + self.num_buckets, self.relative_max_distance, relative_positions, False + ) + + # [batch_size, ngram, sequence_length, hidden_size] + hidden_states = hidden_states.transpose(1, 2) + rel_pos_embeddings = self.relative_pos_embeddings(hidden_states) + + # [batch_size, ngram, sequence_length, num_buckets, num_heads] + rel_pos_embeddings = rel_pos_embeddings.view( + hidden_states.shape[:-1] + (self.num_buckets, self.num_attn_heads) + ) + rel_pos_embeddings = rel_pos_embeddings.permute(0, 2, 1, 4, 3) + # [batch_size * ngram * sequence_length * num_heads, num_buckets] + rel_pos_embeddings = rel_pos_embeddings.reshape(-1, self.num_buckets) + # [ngram, batch_size, num_heads * sequence_length, -1] + predict_relative_position_buckets = predict_relative_position_buckets.unsqueeze(0) + predict_relative_position_buckets = predict_relative_position_buckets.repeat( + self.ngram, 1, self.num_attn_heads, 1 + ) + # [ngram * batch_size * num_heads * sequence_length, -1] + predict_relative_position_buckets = predict_relative_position_buckets.view( + -1, predict_relative_position_buckets.size(-1) + ).long() + + predict_relative_pos_embeddings = torch.gather( + rel_pos_embeddings, dim=1, index=predict_relative_position_buckets + ) + + # [batch_size, gram, num_heads, sequence_length, -1] + predict_relative_pos_embeddings = predict_relative_pos_embeddings.view( + batch_size, self.ngram, self.num_attn_heads, sequence_length, -1 + ) + + return predict_relative_pos_embeddings + + +class ProphetNetEncoderLayer(nn.Module): + """ + Encoder block for Prophetnet + """ + + def __init__(self, config: ProphetNetConfig): + super().__init__() + # 1st residual block + self.self_attn = ProphetNetAttention(config, config.num_encoder_attention_heads) + self.self_attn_layer_norm = LayerNorm(config.hidden_size) + + # 2nd residual block + self.feed_forward = ProphetNetFeedForward(config, config.encoder_ffn_dim) + self.feed_forward_layer_norm = LayerNorm(config.hidden_size) + + def forward( + self, + hidden_states, + attention_mask, + layer_head_mask, + output_attentions: bool = False, + ): + # 1st residual block + attention_output, attn_weights, _ = self.self_attn( + hidden_states=hidden_states, + attention_mask=attention_mask, + layer_head_mask=layer_head_mask, + output_attentions=output_attentions, + ) + hidden_states = self.self_attn_layer_norm(attention_output + hidden_states) + + # 2nd residual block + feed_forward_output = self.feed_forward(hidden_states) + hidden_states = self.feed_forward_layer_norm(feed_forward_output + hidden_states) + + outputs = (hidden_states,) + + if output_attentions: + outputs += (attn_weights,) + + return outputs + + +class ProphetNetDecoderLayer(nn.Module): + """ + Decoder block for Prophetnet + """ + + def __init__(self, config: ProphetNetConfig): + super().__init__() + # 1st residual block + self.self_attn = ProphetNetNgramSelfAttention(config) + self.self_attn_layer_norm = LayerNorm(config.hidden_size) + + # 2nd residual block + if config.add_cross_attention: + self.cross_attn = ProphetNetAttention(config, config.num_decoder_attention_heads) + self.cross_attn_layer_norm = LayerNorm(config.hidden_size) + + # 3rd residual block + self.feed_forward = ProphetNetFeedForward(config, config.decoder_ffn_dim) + self.feed_forward_layer_norm = LayerNorm(config.hidden_size) + + def forward( + self, + hidden_states, + attention_mask=None, + encoder_hidden_states=None, + encoder_attn_mask=None, + layer_head_mask=None, + cross_attn_layer_head_mask=None, + extended_predict_attention_mask=None, + main_relative_position_buckets=None, + predict_relative_position_buckets=None, + position_ids=None, + past_key_value=None, + use_cache: bool = True, + output_attentions: bool = False, + ): + # 1st residual block + # decoder uni-directional self-attention cached key/values tuple is at positions 1,2 + self_attn_past_key_value = past_key_value[:2] if past_key_value is not None else None + ngram_attention_output, self_attn_weights, self_attn_weights_ngram, present_key_value = self.self_attn( + hidden_states=hidden_states, + past_key_value=self_attn_past_key_value, + attention_mask=attention_mask, + layer_head_mask=layer_head_mask, + extended_predict_attention_mask=extended_predict_attention_mask, + main_relative_position_buckets=main_relative_position_buckets, + predict_relative_position_buckets=predict_relative_position_buckets, + position_ids=position_ids, + ) + hidden_states = self.self_attn_layer_norm(hidden_states + ngram_attention_output) + + # cross_attn cached key/values tuple is at positions 3,4 of present_key_value tuple + cross_attn_past_key_value = past_key_value[-2:] if past_key_value is not None else None + cross_attn_weights = None + if encoder_hidden_states is not None: + # 2nd residual block + attention_output, cross_attn_weights, cross_attn_present_key_value = self.cross_attn( + hidden_states=hidden_states, + key_value_states=encoder_hidden_states, + attention_mask=encoder_attn_mask, + layer_head_mask=cross_attn_layer_head_mask, + past_key_value=cross_attn_past_key_value, + output_attentions=output_attentions, + ) + hidden_states = self.cross_attn_layer_norm(attention_output + hidden_states) + + # add cross-attn to positions 3,4 of present_key_value tuple + present_key_value = present_key_value + cross_attn_present_key_value + + # 3rd residual block + feed_forward_output = self.feed_forward(hidden_states) + hidden_states = self.feed_forward_layer_norm(feed_forward_output + hidden_states) + + outputs = (hidden_states,) + + if output_attentions: + outputs += (self_attn_weights, self_attn_weights_ngram, cross_attn_weights) + + if use_cache: + outputs += (present_key_value,) + + return outputs + + +@add_start_docstrings( + "The standalone encoder part of the ProphetNetModel.", + PROPHETNET_START_DOCSTRING, +) +class ProphetNetEncoder(ProphetNetPreTrainedModel): + r""" + word_embeddings (`torch.nn.Embeddings` of shape `(config.vocab_size, config.hidden_size)`, *optional*): + The word embedding parameters. This can be used to initialize [`ProphetNetEncoder`] with pre-defined word + embeddings instead of randomly initialized word embeddings. + """ + + def __init__(self, config: ProphetNetConfig, word_embeddings: nn.Embedding = None): + super().__init__(config) + + self.word_embeddings = ( + word_embeddings + if word_embeddings is not None + else nn.Embedding(config.vocab_size, config.hidden_size, padding_idx=config.pad_token_id) + ) + self.position_embeddings = ProphetNetPositionalEmbeddings(config) + self.embeddings_layer_norm = LayerNorm(config.hidden_size) + + self.layers = nn.ModuleList([ProphetNetEncoderLayer(config) for _ in range(config.num_encoder_layers)]) + + self.gradient_checkpointing = False + # Initialize weights and apply final processing + self.post_init() + + def get_input_embeddings(self): + return self.word_embeddings + + def set_input_embeddings(self, value): + self.word_embeddings = value + + @add_start_docstrings_to_model_forward(PROPHETNET_STANDALONE_INPUTS_DOCSTRING) + @replace_return_docstrings(output_type=BaseModelOutput, config_class=_CONFIG_FOR_DOC) + def forward( + self, + input_ids: Optional[torch.Tensor] = None, + attention_mask: Optional[torch.Tensor] = None, + head_mask: Optional[torch.Tensor] = None, + inputs_embeds: Optional[torch.Tensor] = None, + output_attentions: Optional[bool] = None, + output_hidden_states: Optional[bool] = None, + return_dict: Optional[bool] = None, + ) -> Union[Tuple, BaseModelOutput]: + r""" + Returns: + + Example: + + ```python + >>> from transformers import AutoTokenizer, ProphetNetEncoder + >>> import torch + + >>> tokenizer = AutoTokenizer.from_pretrained("microsoft/prophetnet-large-uncased") + >>> model = ProphetNetEncoder.from_pretrained("patrickvonplaten/prophetnet-large-uncased-standalone") + >>> inputs = tokenizer("Hello, my dog is cute", return_tensors="pt") + >>> outputs = model(**inputs) + + >>> last_hidden_states = outputs.last_hidden_state + ```""" + + output_attentions = output_attentions if output_attentions is not None else self.config.output_attentions + output_hidden_states = ( + output_hidden_states if output_hidden_states is not None else self.config.output_hidden_states + ) + return_dict = return_dict if return_dict is not None else self.config.use_return_dict + + if input_ids is None and inputs_embeds is None: + raise ValueError("Either input_ids or inputs_embeds has to be passed.") + elif input_ids is not None and inputs_embeds is not None: + raise ValueError("Make sure to only pass input_ids or inputs_embeds.") + elif input_ids is not None and inputs_embeds is None: + inputs_embeds = self.word_embeddings(input_ids) + + # prepare attention mask + if attention_mask is not None: + extended_attention_mask = ( + 1.0 - attention_mask[:, None, None, :].repeat(1, self.config.num_encoder_attention_heads, 1, 1) + ) * torch.finfo(self.dtype).min + extended_attention_mask = extended_attention_mask.to(inputs_embeds.dtype) + else: + extended_attention_mask = None + + position_embeddings, position_ids = self.position_embeddings(inputs_embeds.shape[:2], inputs_embeds.device) + + hidden_states = inputs_embeds + position_embeddings + hidden_states = self.embeddings_layer_norm(hidden_states) + hidden_states = nn.functional.dropout(hidden_states, p=self.config.dropout, training=self.training) + + encoder_hidden_states = () if output_hidden_states else None + all_attentions = () if output_attentions else None + + # check if head_mask has a correct number of layers specified if desired + if head_mask is not None: + assert head_mask.size()[0] == ( + len(self.layers) + ), f"The head_mask should be specified for {len(self.layers)} layers, but it is for {head_mask.size()[0]}." + for idx, encoder_layer in enumerate(self.layers): + if output_hidden_states: + encoder_hidden_states = encoder_hidden_states + (hidden_states,) + + if self.gradient_checkpointing and self.training: + layer_outputs = self._gradient_checkpointing_func( + encoder_layer.__call__, + hidden_states, + extended_attention_mask, + (head_mask[idx] if head_mask is not None else None), + output_attentions, + ) + else: + layer_outputs = encoder_layer( + hidden_states, + attention_mask=extended_attention_mask, + layer_head_mask=(head_mask[idx] if head_mask is not None else None), + output_attentions=output_attentions, + ) + + hidden_states = layer_outputs[0] + + if output_attentions: + all_attentions = all_attentions + (layer_outputs[1],) + + if output_hidden_states: + encoder_hidden_states = encoder_hidden_states + (hidden_states,) + + if not return_dict: + return tuple(v for v in [hidden_states, encoder_hidden_states, all_attentions] if v is not None) + return BaseModelOutput( + last_hidden_state=hidden_states, hidden_states=encoder_hidden_states, attentions=all_attentions + ) + + +@add_start_docstrings( + "The standalone decoder part of the ProphetNetModel.", + PROPHETNET_START_DOCSTRING, +) +class ProphetNetDecoder(ProphetNetPreTrainedModel): + r""" + word_embeddings (`torch.nn.Embeddings` of shape `(config.vocab_size, config.hidden_size)`, *optional*): + The word embedding parameters. This can be used to initialize [`ProphetNetEncoder`] with pre-defined word + embeddings instead of randomly initialized word embeddings. + """ + + def __init__(self, config: ProphetNetConfig, word_embeddings: Optional[nn.Embedding] = None): + super().__init__(config) + + self.ngram = config.ngram + self.num_buckets = config.num_buckets + self.relative_max_distance = config.relative_max_distance + self.dropout = config.dropout + self.max_target_positions = config.max_position_embeddings + + self.word_embeddings = ( + word_embeddings + if word_embeddings is not None + else nn.Embedding(config.vocab_size, config.hidden_size, padding_idx=config.pad_token_id) + ) + self.position_embeddings = ProphetNetPositionalEmbeddings(config) + + self.ngram_embeddings = nn.Embedding(self.ngram, config.hidden_size, None) + self.layers = nn.ModuleList([ProphetNetDecoderLayer(config) for _ in range(config.num_decoder_layers)]) + self.embeddings_layer_norm = LayerNorm(config.hidden_size) + + self.gradient_checkpointing = False + # Initialize weights and apply final processing + self.post_init() + + def get_input_embeddings(self): + return self.word_embeddings + + def set_input_embeddings(self, value): + self.word_embeddings = value + + @add_start_docstrings_to_model_forward(PROPHETNET_STANDALONE_INPUTS_DOCSTRING) + @replace_return_docstrings(output_type=ProphetNetDecoderModelOutput, config_class=_CONFIG_FOR_DOC) + def forward( + self, + input_ids: Optional[torch.Tensor] = None, + attention_mask: Optional[torch.Tensor] = None, + encoder_hidden_states: Optional[torch.Tensor] = None, + encoder_attention_mask: Optional[torch.Tensor] = None, + head_mask: Optional[torch.Tensor] = None, + cross_attn_head_mask: Optional[torch.Tensor] = None, + past_key_values: Optional[Tuple[Tuple[torch.Tensor]]] = None, + inputs_embeds: Optional[torch.Tensor] = None, + use_cache: Optional[bool] = None, + output_attentions: Optional[bool] = None, + output_hidden_states: Optional[bool] = None, + return_dict: Optional[bool] = None, + ) -> Union[Tuple, ProphetNetDecoderModelOutput]: + r""" + encoder_hidden_states (`torch.FloatTensor` of shape `(batch_size, sequence_length, hidden_size)`, *optional*): + Sequence of hidden-states at the output of the last layer of the encoder. Used in the cross-attention if + the model is configured as a decoder. + encoder_attention_mask (`torch.FloatTensor` of shape `(batch_size, sequence_length)`, *optional*): + Mask to avoid performing attention on the padding token indices of the encoder input. This mask is used in + the cross-attention if the model is configured as a decoder. Mask values selected in `[0, 1]`: + cross_attn_head_mask (`torch.Tensor` of shape `(decoder_layers, decoder_attention_heads)`, *optional*): + Mask to nullify selected heads of the cross-attention modules. Mask values selected in `[0, 1]`: + + - 1 indicates the head is **not masked**, + - 0 indicates the head is **masked**. + + past_key_values (`tuple(tuple(torch.FloatTensor))` of length `config.n_layers` with each tuple having 4 tensors of shape `(batch_size, num_heads, sequence_length - 1, embed_size_per_head)`): + Contains precomputed key and value hidden-states of the attention blocks. Can be used to speed up decoding. + + If `past_key_values` are used, the user can optionally input only the last `decoder_input_ids` (those that + don't have their past key value states given to this model) of shape `(batch_size, 1)` instead of all + `decoder_input_ids` of shape `(batch_size, sequence_length)`. + use_cache (`bool`, *optional*): + If set to `True`, `past_key_values` key value states are returned and can be used to speed up decoding (see + `past_key_values`). + + - 1 for tokens that are **not masked**, + - 0 for tokens that are **masked**. + + Returns: + + Example: + + ```python + >>> from transformers import AutoTokenizer, ProphetNetDecoder + >>> import torch + + >>> tokenizer = AutoTokenizer.from_pretrained("microsoft/prophetnet-large-uncased") + >>> model = ProphetNetDecoder.from_pretrained("microsoft/prophetnet-large-uncased", add_cross_attention=False) + >>> inputs = tokenizer("Hello, my dog is cute", return_tensors="pt") + >>> outputs = model(**inputs) + + >>> last_hidden_states = outputs.last_hidden_state + ```""" + use_cache = use_cache if use_cache is not None else self.config.use_cache + output_attentions = output_attentions if output_attentions is not None else self.config.output_attentions + output_hidden_states = ( + output_hidden_states if output_hidden_states is not None else self.config.output_hidden_states + ) + return_dict = return_dict if return_dict is not None else self.config.use_return_dict + + if input_ids is None and inputs_embeds is None: + raise ValueError("Either `decoder_input_ids` or `decoder_inputs_embeds` has to be passed.") + elif input_ids is not None and inputs_embeds is not None: + raise ValueError("Make sure to only pass `decoder_input_ids` or `decoder_inputs_embeds`.") + elif input_ids is not None and inputs_embeds is None: + inputs_embeds = self.word_embeddings(input_ids) + + batch_size, sequence_length = inputs_embeds.shape[:2] + + main_stream_pos_embed, position_ids = self.position_embeddings( + (batch_size, sequence_length), + device=inputs_embeds.device, + past_key_values=past_key_values, + ) + + if past_key_values is not None: + main_relative_position_buckets, predict_relative_position_buckets = None, None + else: + ( + main_relative_position_buckets, + predict_relative_position_buckets, + ) = self.compute_buffered_relative_buckets(position_ids) + predicting_stream_pos_embed = self.position_embeddings._forward(position_ids + 1) + + # add position embeddings + hidden_states = inputs_embeds + main_stream_pos_embed + + ngram_embeddings = self.ngram_embeddings.weight + + # prepare attention mask + if past_key_values is not None: + assert ( + hidden_states.size(1) == 1 + ), "At the moment `use_cache` is only supported for `decoder_input_ids` of length 1" + + ngram_hidden_states = [ + (ngram_embeddings[ngram - 1] + predicting_stream_pos_embed).repeat(batch_size, 1, 1) + for ngram in range(self.ngram) + ] + extended_attention_mask = None + extended_predict_attention_mask = None + else: + ngram_hidden_states = [ + (ngram_embeddings[ngram - 1] + predicting_stream_pos_embed) for ngram in range(self.ngram) + ] + extended_attention_mask = self.prepare_attention_mask(hidden_states, attention_mask) + extended_predict_attention_mask = self.prepare_predict_attention_mask(hidden_states, attention_mask) + + # prepare encoder attention mask + if encoder_attention_mask is not None: + extended_encoder_attention_mask = ( + 1.0 - encoder_attention_mask[:, None, None, :].repeat(1, self.config.num_decoder_attention_heads, 1, 1) + ) * torch.finfo(self.dtype).min + extended_encoder_attention_mask = extended_encoder_attention_mask.to(inputs_embeds.dtype) + else: + extended_encoder_attention_mask = None + + hidden_states = torch.cat([hidden_states] + ngram_hidden_states, 1) + + if self.embeddings_layer_norm: + hidden_states = self.embeddings_layer_norm(hidden_states) + + hidden_states = nn.functional.dropout(hidden_states, p=self.dropout, training=self.training) + + # init attentions, hidden_states and cache with empty tuples + all_main_stream_hidden_states = () if output_hidden_states else None + all_ngram_stream_hidden_states = () if output_hidden_states and self.config.ngram > 0 else None + + all_main_stream_attns = () if output_attentions else None + all_ngram_stream_attns = () if output_attentions else None + all_cross_attns = () if output_attentions and self.config.add_cross_attention else None + + if self.gradient_checkpointing and self.training: + if use_cache: + logger.warning_once( + "`use_cache=True` is incompatible with gradient checkpointing. Setting `use_cache=False`..." + ) + use_cache = False + + present_key_values = () if use_cache else None + + # check if head_mask/cross_attn_head_mask has a correct number of layers specified if desired + for attn_mask, mask_name in zip([head_mask, cross_attn_head_mask], ["head_mask", "cross_attn_head_mask"]): + if attn_mask is not None: + assert attn_mask.size()[0] == (len(self.layers)), ( + f"The `{mask_name}` should be specified for {len(self.layers)} layers, but it is for" + f" {head_mask.size()[0]}." + ) + for idx, decoder_layer in enumerate(self.layers): + if output_hidden_states: + # grad cannot be kept because tensor is sliced + all_main_stream_hidden_states += (hidden_states[:, :sequence_length],) + if self.config.ngram > 0: + all_ngram_stream_hidden_states += (hidden_states[:, sequence_length:],) + + past_key_value = past_key_values[idx] if past_key_values is not None else None + + if self.gradient_checkpointing and self.training: + layer_outputs = self._gradient_checkpointing_func( + decoder_layer.__call__, + hidden_states, + extended_attention_mask, + encoder_hidden_states, + extended_encoder_attention_mask, + (head_mask[idx] if head_mask is not None else None), + (cross_attn_head_mask[idx] if cross_attn_head_mask is not None else None), + extended_predict_attention_mask, + main_relative_position_buckets, + predict_relative_position_buckets, + position_ids, + None, + use_cache, + output_attentions, + ) + else: + layer_outputs = decoder_layer( + hidden_states, + attention_mask=extended_attention_mask, + encoder_hidden_states=encoder_hidden_states, + encoder_attn_mask=extended_encoder_attention_mask, + layer_head_mask=(head_mask[idx] if head_mask is not None else None), + cross_attn_layer_head_mask=( + cross_attn_head_mask[idx] if cross_attn_head_mask is not None else None + ), + extended_predict_attention_mask=extended_predict_attention_mask, + main_relative_position_buckets=main_relative_position_buckets, + predict_relative_position_buckets=predict_relative_position_buckets, + position_ids=position_ids, + past_key_value=past_key_value, + use_cache=use_cache, + output_attentions=output_attentions, + ) + + hidden_states = layer_outputs[0] + + if use_cache: + present_key_values += (layer_outputs[4 if output_attentions else 1],) + + if output_attentions: + all_main_stream_attns += (layer_outputs[1],) + all_ngram_stream_attns += (layer_outputs[2],) + + if self.config.add_cross_attention: + all_cross_attns += (layer_outputs[3],) + + if output_hidden_states: + all_main_stream_hidden_states += (hidden_states[:, :sequence_length],) + if self.config.ngram > 0: + all_ngram_stream_hidden_states += (hidden_states[:, sequence_length:],) + + # split last_hidden_state for return + last_hidden_state = hidden_states[:, :sequence_length] + last_hidden_state_ngram = hidden_states[:, sequence_length:] if self.config.ngram > 0 else None + + if not return_dict: + return tuple( + v + for v in [ + last_hidden_state, + last_hidden_state_ngram, + present_key_values, + all_main_stream_hidden_states, + all_ngram_stream_hidden_states, + all_main_stream_attns, + all_ngram_stream_attns, + all_cross_attns, + ] + if v is not None + ) + return ProphetNetDecoderModelOutput( + last_hidden_state=last_hidden_state, + last_hidden_state_ngram=last_hidden_state_ngram, + past_key_values=present_key_values, + hidden_states=all_main_stream_hidden_states, + hidden_states_ngram=all_ngram_stream_hidden_states, + attentions=all_main_stream_attns, + ngram_attentions=all_ngram_stream_attns, + cross_attentions=all_cross_attns, + ) + + def compute_buffered_relative_buckets(self, position_ids): + batch_size, sequence_length = position_ids.shape + + position_ids = torch.arange(1, self.max_target_positions).to(position_ids.device).repeat(1, 1) + main_relative_buckets, predict_relative_buckets = compute_all_stream_relative_buckets( + self.num_buckets, self.relative_max_distance, position_ids + ) + + # buffer relative buckets + main_relative_buckets = main_relative_buckets[:, :sequence_length, :sequence_length].repeat(batch_size, 1, 1) + predict_relative_buckets = torch.cat( + [ + predict_relative_buckets[:, :sequence_length, :sequence_length], + predict_relative_buckets[ + :, :sequence_length, self.max_target_positions : self.max_target_positions + sequence_length + ], + ], + 2, + ).repeat(batch_size, 1, 1) + + return main_relative_buckets, predict_relative_buckets + + def prepare_attention_mask(self, hidden_states, attention_mask): + batch_size, seq_length = hidden_states.shape[:2] + + # get causal mask + causal_mask = torch.full( + (seq_length, seq_length), + torch.finfo(hidden_states.dtype).min, + dtype=hidden_states.dtype, + device=hidden_states.device, + ) + causal_mask = torch.triu(causal_mask, 1) + + extended_causal_mask = causal_mask[:seq_length, :seq_length][None, None, :, :].expand( + (batch_size, self.config.num_decoder_attention_heads) + causal_mask.shape + ) + + # add usual attention mask + if attention_mask is not None: + extended_attention_mask = (1.0 - attention_mask[:, None, None, :]) * torch.finfo(self.dtype).min + extended_attention_mask = extended_causal_mask + extended_attention_mask + else: + extended_attention_mask = extended_causal_mask + return extended_attention_mask.to(hidden_states.dtype) + + def prepare_predict_attention_mask(self, hidden_states, attention_mask): + batch_size, seq_length = hidden_states.shape[:2] + + # get causal mask + predict_causal_mask = ngram_attention_bias( + self.max_target_positions, self.ngram, hidden_states.device, hidden_states.dtype + ) + predict_causal_mask = torch.cat( + [ + predict_causal_mask[:, :seq_length, :seq_length], + predict_causal_mask[ + :, :seq_length, self.max_target_positions : self.max_target_positions + seq_length + ], + ], + dim=-1, + ) + extended_predict_causal_mask = predict_causal_mask[None, None, :, :, :].expand( + (batch_size, self.config.num_decoder_attention_heads) + predict_causal_mask.shape + ) + + # add usual attention mask + if attention_mask is not None: + extended_attention_mask = (1.0 - attention_mask[:, None, None, None, :]) * torch.finfo(self.dtype).min + extended_attention_mask = extended_attention_mask.expand( + (batch_size, self.config.num_decoder_attention_heads, self.ngram, seq_length, seq_length) + ) + # predicted stream attention_mask should always be 0 + extended_attention_mask = torch.cat( + [extended_attention_mask, torch.zeros_like(extended_attention_mask)], dim=-1 + ) + extended_predict_attention_mask = extended_predict_causal_mask + extended_attention_mask + else: + extended_predict_attention_mask = extended_predict_causal_mask + return extended_predict_attention_mask.to(hidden_states.dtype) + + +@add_start_docstrings( + "The bare ProphetNet Model outputting raw hidden-states without any specific head on top.", + PROPHETNET_START_DOCSTRING, +) +class ProphetNetModel(ProphetNetPreTrainedModel): + _tied_weights_keys = ["encoder.word_embeddings.weight", "decoder.word_embeddings.weight"] + + def __init__(self, config: ProphetNetConfig): + super().__init__(config) + self.word_embeddings = nn.Embedding(config.vocab_size, config.hidden_size, padding_idx=config.pad_token_id) + + encoder_config = copy.deepcopy(config) + encoder_config.is_encoder_decoder = False + encoder_config.use_cache = False + self.encoder = ProphetNetEncoder(encoder_config, self.word_embeddings) + + decoder_config = copy.deepcopy(config) + decoder_config.is_decoder = True + decoder_config.is_encoder_decoder = False + self.decoder = ProphetNetDecoder(decoder_config, self.word_embeddings) + + # Initialize weights and apply final processing + self.post_init() + + def get_input_embeddings(self): + return self.word_embeddings + + def set_input_embeddings(self, value): + self.word_embeddings = value + self.encoder.word_embeddings = self.word_embeddings + self.decoder.word_embeddings = self.word_embeddings + + def _tie_weights(self): + if self.config.tie_word_embeddings: + self._tie_or_clone_weights(self.encoder.word_embeddings, self.word_embeddings) + self._tie_or_clone_weights(self.decoder.word_embeddings, self.word_embeddings) + + def get_encoder(self): + return self.encoder + + def get_decoder(self): + return self.decoder + + @add_start_docstrings_to_model_forward(PROPHETNET_INPUTS_DOCSTRING) + @replace_return_docstrings(output_type=ProphetNetSeq2SeqModelOutput, config_class=_CONFIG_FOR_DOC) + def forward( + self, + input_ids: Optional[torch.Tensor] = None, + attention_mask: Optional[torch.Tensor] = None, + decoder_input_ids: Optional[torch.Tensor] = None, + decoder_attention_mask: Optional[torch.BoolTensor] = None, + head_mask: Optional[torch.Tensor] = None, + decoder_head_mask: Optional[torch.Tensor] = None, + cross_attn_head_mask: Optional[torch.Tensor] = None, + encoder_outputs: Optional[Tuple] = None, + past_key_values: Optional[Tuple[Tuple[torch.Tensor]]] = None, + inputs_embeds: Optional[torch.Tensor] = None, + decoder_inputs_embeds: Optional[torch.Tensor] = None, + use_cache: Optional[bool] = None, + output_attentions: Optional[bool] = None, + output_hidden_states: Optional[bool] = None, + return_dict: Optional[bool] = None, + ) -> Union[Tuple, ProphetNetSeq2SeqModelOutput]: + r""" + Returns: + + Example: + + ```python + >>> from transformers import AutoTokenizer, ProphetNetModel + + >>> tokenizer = AutoTokenizer.from_pretrained("microsoft/prophetnet-large-uncased") + >>> model = ProphetNetModel.from_pretrained("microsoft/prophetnet-large-uncased") + + >>> input_ids = tokenizer( + ... "Studies have been shown that owning a dog is good for you", return_tensors="pt" + ... ).input_ids # Batch size 1 + >>> decoder_input_ids = tokenizer("Studies show that", return_tensors="pt").input_ids # Batch size 1 + >>> outputs = model(input_ids=input_ids, decoder_input_ids=decoder_input_ids) + + >>> last_hidden_states = outputs.last_hidden_state # main stream hidden states + >>> last_hidden_states_ngram = outputs.last_hidden_state_ngram # predict hidden states + ```""" + use_cache = use_cache if use_cache is not None else self.config.use_cache + output_attentions = output_attentions if output_attentions is not None else self.config.output_attentions + output_hidden_states = ( + output_hidden_states if output_hidden_states is not None else self.config.output_hidden_states + ) + return_dict = return_dict if return_dict is not None else self.config.use_return_dict + + if encoder_outputs is None: + encoder_outputs = self.encoder( + input_ids=input_ids, + attention_mask=attention_mask, + head_mask=head_mask, + inputs_embeds=inputs_embeds, + output_attentions=output_attentions, + output_hidden_states=output_hidden_states, + return_dict=return_dict, + ) + + # decoder outputs consists of (dec_features, past_key_values, dec_hidden, dec_attn) + decoder_outputs = self.decoder( + input_ids=decoder_input_ids, + attention_mask=decoder_attention_mask, + encoder_hidden_states=encoder_outputs[0], + encoder_attention_mask=attention_mask, + head_mask=decoder_head_mask, + cross_attn_head_mask=cross_attn_head_mask, + past_key_values=past_key_values, + inputs_embeds=decoder_inputs_embeds, + output_attentions=output_attentions, + output_hidden_states=output_hidden_states, + use_cache=use_cache, + return_dict=return_dict, + ) + + if not return_dict: + return decoder_outputs + encoder_outputs + return ProphetNetSeq2SeqModelOutput( + last_hidden_state=decoder_outputs.last_hidden_state, + last_hidden_state_ngram=decoder_outputs.last_hidden_state_ngram, + past_key_values=decoder_outputs.past_key_values, + decoder_hidden_states=decoder_outputs.hidden_states, + decoder_ngram_hidden_states=decoder_outputs.hidden_states_ngram, + decoder_attentions=decoder_outputs.attentions, + decoder_ngram_attentions=decoder_outputs.ngram_attentions, + cross_attentions=decoder_outputs.cross_attentions, + encoder_last_hidden_state=encoder_outputs.last_hidden_state, + encoder_hidden_states=encoder_outputs.hidden_states, + encoder_attentions=encoder_outputs.attentions, + ) + + +@add_start_docstrings( + "The ProphetNet Model with a language modeling head. Can be used for sequence generation tasks.", + PROPHETNET_START_DOCSTRING, +) +class ProphetNetForConditionalGeneration(ProphetNetPreTrainedModel): + _tied_weights_keys = ["encoder.word_embeddings.weight", "decoder.word_embeddings.weight", "lm_head.weight"] + + def __init__(self, config: ProphetNetConfig): + super().__init__(config) + self.prophetnet = ProphetNetModel(config) + self.padding_idx = config.pad_token_id + self.disable_ngram_loss = config.disable_ngram_loss + + self.lm_head = nn.Linear(config.hidden_size, config.vocab_size, bias=False) + + # Initialize weights and apply final processing + self.post_init() + + def get_output_embeddings(self): + return self.lm_head + + def set_output_embeddings(self, new_embeddings): + self.lm_head = new_embeddings + + def _tie_weights(self): + if self.config.tie_word_embeddings: + self._tie_or_clone_weights(self.prophetnet.word_embeddings, self.lm_head) + + def get_input_embeddings(self): + return self.prophetnet.word_embeddings + + @add_start_docstrings_to_model_forward(PROPHETNET_INPUTS_DOCSTRING) + @replace_return_docstrings(output_type=ProphetNetSeq2SeqLMOutput, config_class=_CONFIG_FOR_DOC) + def forward( + self, + input_ids: Optional[torch.Tensor] = None, + attention_mask: Optional[torch.Tensor] = None, + decoder_input_ids: Optional[torch.Tensor] = None, + decoder_attention_mask: Optional[torch.BoolTensor] = None, + head_mask: Optional[torch.Tensor] = None, + decoder_head_mask: Optional[torch.Tensor] = None, + cross_attn_head_mask: Optional[torch.Tensor] = None, + encoder_outputs: Optional[torch.Tensor] = None, + past_key_values: Optional[Tuple[Tuple[torch.Tensor]]] = None, + inputs_embeds: Optional[torch.Tensor] = None, + decoder_inputs_embeds: Optional[torch.Tensor] = None, + labels: Optional[torch.Tensor] = None, + use_cache: Optional[bool] = None, + output_attentions: Optional[bool] = None, + output_hidden_states: Optional[bool] = None, + return_dict: Optional[bool] = None, + ) -> Union[Tuple, ProphetNetSeq2SeqLMOutput]: + r""" + labels (`torch.LongTensor` of shape `(batch_size,)`, *optional*): + Labels for computing the sequence classification/regression loss. Indices should be in `[-100, 0, ..., + config.vocab_size - 1]`. All labels set to `-100` are ignored (masked), the loss is only computed for + labels in `[0, ..., config.vocab_size]` + + Returns: + + Example: + + ```python + >>> from transformers import AutoTokenizer, ProphetNetForConditionalGeneration + + >>> tokenizer = AutoTokenizer.from_pretrained("microsoft/prophetnet-large-uncased") + >>> model = ProphetNetForConditionalGeneration.from_pretrained("microsoft/prophetnet-large-uncased") + + >>> input_ids = tokenizer( + ... "Studies have been shown that owning a dog is good for you", return_tensors="pt" + ... ).input_ids # Batch size 1 + >>> decoder_input_ids = tokenizer("Studies show that", return_tensors="pt").input_ids # Batch size 1 + >>> outputs = model(input_ids=input_ids, decoder_input_ids=decoder_input_ids) + + >>> logits_next_token = outputs.logits # logits to predict next token as usual + >>> logits_ngram_next_tokens = outputs.logits_ngram # logits to predict 2nd, 3rd, ... next tokens + ```""" + return_dict = return_dict if return_dict is not None else self.config.use_return_dict + + if labels is not None and decoder_input_ids is None and decoder_inputs_embeds is None: + # get decoder inputs from shifting lm labels to the right + decoder_input_ids = self._shift_right(labels) + + outputs = self.prophetnet( + input_ids=input_ids, + attention_mask=attention_mask, + decoder_input_ids=decoder_input_ids, + decoder_attention_mask=decoder_attention_mask, + head_mask=head_mask, + decoder_head_mask=decoder_head_mask, + cross_attn_head_mask=cross_attn_head_mask, + encoder_outputs=encoder_outputs, + past_key_values=past_key_values, + inputs_embeds=inputs_embeds, + decoder_inputs_embeds=decoder_inputs_embeds, + use_cache=use_cache, + output_attentions=output_attentions, + output_hidden_states=output_hidden_states, + return_dict=return_dict, + ) + batch_size, sequence_length = ( + decoder_input_ids.shape if decoder_input_ids is not None else decoder_inputs_embeds.shape[:2] + ) + + predicting_streams = outputs[1].view(batch_size, self.config.ngram, sequence_length, -1) + predict_logits = self.lm_head(predicting_streams) + + logits = predict_logits[:, 0] + logits_ngram = predict_logits[:, 1:] if self.config.ngram > 1 else None + + # To use .view in loss computation, make sure that logits is contiguous. + if not logits.is_contiguous(): + logits = logits.contiguous() + + loss = None + if labels is not None: + loss = self._compute_loss(predict_logits, labels) + + if not return_dict: + all_logits = tuple(v for v in [logits, logits_ngram] if v is not None) + return (loss,) + all_logits + outputs[2:] if loss is not None else all_logits + outputs[2:] + else: + return ProphetNetSeq2SeqLMOutput( + loss=loss, + logits=logits, + logits_ngram=logits_ngram, + past_key_values=outputs.past_key_values, + decoder_hidden_states=outputs.decoder_hidden_states, + decoder_ngram_hidden_states=outputs.decoder_ngram_hidden_states, + decoder_attentions=outputs.decoder_attentions, + decoder_ngram_attentions=outputs.decoder_ngram_attentions, + cross_attentions=outputs.cross_attentions, + encoder_last_hidden_state=outputs.encoder_last_hidden_state, + encoder_hidden_states=outputs.encoder_hidden_states, + encoder_attentions=outputs.encoder_attentions, + ) + + def _compute_loss(self, logits, labels, ignore_index=-100): + expend_targets = labels.new_zeros(self.config.ngram, labels.size(0), labels.size(1)).fill_(ignore_index) + + for i in range(self.config.ngram): + if i > 0 and self.disable_ngram_loss: + break + expend_targets[i, :, :] = labels + + logits = logits.transpose(0, 1).contiguous() + lprobs = nn.functional.log_softmax( + logits.view(-1, logits.size(-1)), + dim=-1, + dtype=torch.float32, + ) + + loss = nn.functional.nll_loss(lprobs, expend_targets.view(-1), reduction="mean") + + if self.config.eps > 0.0: + smooth_loss = -lprobs.sum(dim=-1, keepdim=True) + non_masked_tokens = expend_targets.ne(ignore_index).view(-1) + smooth_loss = smooth_loss[non_masked_tokens] + smooth_loss = smooth_loss.mean() + + eps_i = self.config.eps / lprobs.size(-1) + loss = (1.0 - self.config.eps) * loss + eps_i * smooth_loss + + return loss + + def prepare_inputs_for_generation( + self, + decoder_input_ids, + past_key_values=None, + attention_mask=None, + head_mask=None, + decoder_head_mask=None, + cross_attn_head_mask=None, + use_cache=None, + encoder_outputs=None, + **kwargs, + ): + assert encoder_outputs is not None, "`encoder_outputs` have to be passed for generation." + + if past_key_values: + decoder_input_ids = decoder_input_ids[:, -1:] + # first step, decoder_cached_states are empty + return { + "input_ids": None, # encoder_outputs is defined. input_ids not needed + "encoder_outputs": encoder_outputs, + "past_key_values": past_key_values, + "decoder_input_ids": decoder_input_ids, + "attention_mask": attention_mask, + "head_mask": head_mask, + "decoder_head_mask": decoder_head_mask, + "cross_attn_head_mask": cross_attn_head_mask, + "use_cache": use_cache, + } + + def prepare_decoder_input_ids_from_labels(self, labels: torch.Tensor): + return self._shift_right(labels) + + @staticmethod + # Copied from transformers.models.bart.modeling_bart.BartForConditionalGeneration._reorder_cache + def _reorder_cache(past_key_values, beam_idx): + reordered_past = () + for layer_past in past_key_values: + # cached cross_attention states don't have to be reordered -> they are always the same + reordered_past += ( + tuple(past_state.index_select(0, beam_idx.to(past_state.device)) for past_state in layer_past[:2]) + + layer_past[2:], + ) + return reordered_past + + def get_encoder(self): + return self.prophetnet.encoder + + def get_decoder(self): + return self.prophetnet.decoder + + +@add_start_docstrings( + "The standalone decoder part of the ProphetNetModel with a lm head on top. The model can be used for causal" + " language modeling.", + PROPHETNET_START_DOCSTRING, +) +class ProphetNetForCausalLM(ProphetNetPreTrainedModel): + _tied_weights_keys = [ + "prophetnet.word_embeddings.weight", + "prophetnet.decoder.word_embeddings.weight", + "lm_head.weight", + ] + + def __init__(self, config: ProphetNetConfig): + # set config for CLM + config = copy.deepcopy(config) + config.is_decoder = True + config.is_encoder_decoder = False + super().__init__(config) + self.prophetnet = ProphetNetDecoderWrapper(config) + + self.padding_idx = config.pad_token_id + self.disable_ngram_loss = config.disable_ngram_loss + + self.lm_head = nn.Linear(config.hidden_size, config.vocab_size, bias=False) + + # Initialize weights and apply final processing + self.post_init() + + def get_input_embeddings(self): + return self.prophetnet.decoder.word_embeddings + + def set_input_embeddings(self, value): + self.prophetnet.decoder.word_embeddings = value + + def get_output_embeddings(self): + return self.lm_head + + def set_output_embeddings(self, new_embeddings): + self.lm_head = new_embeddings + + def _tie_weights(self): + if self.config.tie_word_embeddings: + self._tie_or_clone_weights(self.prophetnet.decoder.word_embeddings, self.lm_head) + + def set_decoder(self, decoder): + self.prophetnet.decoder = decoder + + def get_decoder(self): + return self.prophetnet.decoder + + @add_start_docstrings_to_model_forward(PROPHETNET_STANDALONE_INPUTS_DOCSTRING) + @replace_return_docstrings(output_type=ProphetNetDecoderLMOutput, config_class=_CONFIG_FOR_DOC) + def forward( + self, + input_ids: Optional[torch.Tensor] = None, + attention_mask: Optional[torch.Tensor] = None, + encoder_hidden_states: Optional[torch.Tensor] = None, + encoder_attention_mask: Optional[torch.Tensor] = None, + head_mask: Optional[torch.Tensor] = None, + cross_attn_head_mask: Optional[torch.Tensor] = None, + past_key_values: Optional[Tuple[Tuple[torch.Tensor]]] = None, + inputs_embeds: Optional[torch.Tensor] = None, + labels: Optional[torch.Tensor] = None, + use_cache: Optional[bool] = None, + output_attentions: Optional[bool] = None, + output_hidden_states: Optional[bool] = None, + return_dict: Optional[bool] = None, + ) -> Union[Tuple, ProphetNetDecoderLMOutput]: + r""" + encoder_hidden_states (`torch.FloatTensor` of shape `(batch_size, sequence_length, hidden_size)`, *optional*): + Sequence of hidden-states at the output of the last layer of the encoder. Used in the cross-attention if + the model is configured as a decoder. + encoder_attention_mask (`torch.FloatTensor` of shape `(batch_size, sequence_length)`, *optional*): + Mask to avoid performing attention on the padding token indices of the encoder input. This mask is used in + the cross-attention if the model is configured as a decoder. Mask values selected in `[0, 1]`: + cross_attn_head_mask (`torch.Tensor` of shape `(decoder_layers, decoder_attention_heads)`, *optional*): + Mask to nullify selected heads of the cross-attention modules. Mask values selected in `[0, 1]`: + + - 1 indicates the head is **not masked**, + - 0 indicates the head is **masked**. + + past_key_values (`tuple(tuple(torch.FloatTensor))` of length `config.n_layers` with each tuple having 4 tensors of shape `(batch_size, num_heads, sequence_length - 1, embed_size_per_head)`): + Contains precomputed key and value hidden-states of the attention blocks. Can be used to speed up decoding. + + If `past_key_values` are used, the user can optionally input only the last `decoder_input_ids` (those that + don't have their past key value states given to this model) of shape `(batch_size, 1)` instead of all + `decoder_input_ids` of shape `(batch_size, sequence_length)`. + use_cache (`bool`, *optional*): + If set to `True`, `past_key_values` key value states are returned and can be used to speed up decoding (see + `past_key_values`). + + - 1 for tokens that are **not masked**, + - 0 for tokens that are **masked**. + + labels (`torch.LongTensor` of shape `(batch_size, sequence_length)`, *optional*): + Labels for computing the left-to-right language modeling loss (next word prediction). Indices should be in + `[-100, 0, ..., config.vocab_size]` (see `input_ids` docstring) Tokens with indices set to `-100` are + ignored (masked), the loss is only computed for the tokens with labels n `[0, ..., config.vocab_size]` + + Returns: + + Example: + + ```python + >>> from transformers import AutoTokenizer, ProphetNetForCausalLM + >>> import torch + + >>> tokenizer = AutoTokenizer.from_pretrained("microsoft/prophetnet-large-uncased") + >>> model = ProphetNetForCausalLM.from_pretrained("microsoft/prophetnet-large-uncased") + >>> assert model.config.is_decoder, f"{model.__class__} has to be configured as a decoder." + >>> inputs = tokenizer("Hello, my dog is cute", return_tensors="pt") + >>> outputs = model(**inputs) + + >>> logits = outputs.logits + + >>> # Model can also be used with EncoderDecoder framework + >>> from transformers import BertTokenizer, EncoderDecoderModel, AutoTokenizer + >>> import torch + + >>> tokenizer_enc = BertTokenizer.from_pretrained("google-bert/bert-large-uncased") + >>> tokenizer_dec = AutoTokenizer.from_pretrained("microsoft/prophetnet-large-uncased") + >>> model = EncoderDecoderModel.from_encoder_decoder_pretrained( + ... "google-bert/bert-large-uncased", "microsoft/prophetnet-large-uncased" + ... ) + + >>> ARTICLE = ( + ... "the us state department said wednesday it had received no " + ... "formal word from bolivia that it was expelling the us ambassador there " + ... "but said the charges made against him are `` baseless ." + ... ) + >>> input_ids = tokenizer_enc(ARTICLE, return_tensors="pt").input_ids + >>> labels = tokenizer_dec( + ... "us rejects charges against its ambassador in bolivia", return_tensors="pt" + ... ).input_ids + >>> outputs = model(input_ids=input_ids, decoder_input_ids=labels[:, :-1], labels=labels[:, 1:]) + + >>> loss = outputs.loss + ```""" + return_dict = return_dict if return_dict is not None else self.config.use_return_dict + + # decoder outputs consists of (dec_features, past_key_values, dec_hidden, dec_attn) + outputs = self.prophetnet.decoder( + input_ids=input_ids, + attention_mask=attention_mask, + encoder_hidden_states=encoder_hidden_states, + encoder_attention_mask=encoder_attention_mask, + head_mask=head_mask, + cross_attn_head_mask=cross_attn_head_mask, + past_key_values=past_key_values, + inputs_embeds=inputs_embeds, + use_cache=use_cache, + output_attentions=output_attentions, + output_hidden_states=output_hidden_states, + return_dict=return_dict, + ) + + batch_size, sequence_length = input_ids.shape if input_ids is not None else inputs_embeds.shape[:2] + + predicting_streams = outputs[1].view(batch_size, self.config.ngram, sequence_length, -1) + predict_logits = self.lm_head(predicting_streams) + + logits = predict_logits[:, 0] + logits_ngram = predict_logits[:, 1:] if self.config.ngram > 1 else None + + loss = None + if labels is not None: + loss = self._compute_loss(predict_logits, labels) + + if not return_dict: + all_logits = tuple(v for v in [logits, logits_ngram] if v is not None) + return (loss,) + all_logits + outputs[2:] if loss is not None else all_logits + outputs[2:] + else: + return ProphetNetDecoderLMOutput( + loss=loss, + logits=logits, + logits_ngram=logits_ngram, + past_key_values=outputs.past_key_values, + hidden_states=outputs.hidden_states, + hidden_states_ngram=outputs.hidden_states_ngram, + attentions=outputs.attentions, + ngram_attentions=outputs.ngram_attentions, + cross_attentions=outputs.cross_attentions, + ) + + def _compute_loss(self, logits, labels, ignore_index=-100): + expend_targets = labels.new_zeros(self.config.ngram, labels.size(0), labels.size(1)).fill_(ignore_index) + + for i in range(self.config.ngram): + if i > 0 and self.disable_ngram_loss: + break + expend_targets[i, :, :] = labels + + logits = logits.transpose(0, 1).contiguous() + lprobs = nn.functional.log_softmax( + logits.view(-1, logits.size(-1)), + dim=-1, + dtype=torch.float32, + ) + + loss = nn.functional.nll_loss(lprobs, expend_targets.view(-1), reduction="mean") + + if self.config.eps > 0.0: + smooth_loss = -lprobs.sum(dim=-1, keepdim=True) + non_masked_tokens = expend_targets.ne(ignore_index).view(-1) + smooth_loss = smooth_loss[non_masked_tokens] + smooth_loss = smooth_loss.mean() + + eps_i = self.config.eps / lprobs.size(-1) + loss = (1.0 - self.config.eps) * loss + eps_i * smooth_loss + + return loss + + def prepare_inputs_for_generation( + self, + input_ids, + past_key_values=None, + attention_mask=None, + head_mask=None, + use_cache=None, + **kwargs, + ): + # if model is used as a decoder in encoder-decoder model, the decoder attention mask is created on the fly + if attention_mask is None: + attention_mask = input_ids.new_ones(input_ids.shape) + + if past_key_values: + input_ids = input_ids[:, -1:] + # first step, decoder_cached_states are empty + return { + "input_ids": input_ids, # encoder_outputs is defined. input_ids not needed + "attention_mask": attention_mask, + "head_mask": head_mask, + "past_key_values": past_key_values, + "use_cache": use_cache, + } + + @staticmethod + # Copied from transformers.models.bart.modeling_bart.BartForCausalLM._reorder_cache + def _reorder_cache(past_key_values, beam_idx): + reordered_past = () + for layer_past in past_key_values: + reordered_past += ( + tuple(past_state.index_select(0, beam_idx.to(past_state.device)) for past_state in layer_past), + ) + return reordered_past + + +class ProphetNetDecoderWrapper(ProphetNetPreTrainedModel): + """ + This is a wrapper class, so that [`ProphetNetForCausalLM`] can correctly be loaded from pretrained prophetnet + classes. + """ + + def __init__(self, config: ProphetNetConfig): + super().__init__(config) + + self.word_embeddings = nn.Embedding(config.vocab_size, config.hidden_size, padding_idx=config.pad_token_id) + self.decoder = ProphetNetDecoder(config, word_embeddings=self.word_embeddings) + + # Initialize weights and apply final processing + self.post_init() + + def _tie_weights(self): + self._tie_or_clone_weights(self.word_embeddings, self.decoder.get_input_embeddings()) + + def forward(self, *args, **kwargs): + return self.decoder(*args, **kwargs) diff --git a/venv/lib/python3.10/site-packages/transformers/models/prophetnet/tokenization_prophetnet.py b/venv/lib/python3.10/site-packages/transformers/models/prophetnet/tokenization_prophetnet.py new file mode 100644 index 0000000000000000000000000000000000000000..cd387520af18efcce7e49fe3450485fd56a0e204 --- /dev/null +++ b/venv/lib/python3.10/site-packages/transformers/models/prophetnet/tokenization_prophetnet.py @@ -0,0 +1,499 @@ +# coding=utf-8 +# Copyright 2020 The Microsoft Authors and The HuggingFace Inc. team. +# +# Licensed under the Apache License, Version 2.0 (the "License"); +# you may not use this file except in compliance with the License. +# You may obtain a copy of the License at +# +# http://www.apache.org/licenses/LICENSE-2.0 +# +# Unless required by applicable law or agreed to in writing, software +# distributed under the License is distributed on an "AS IS" BASIS, +# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. +# See the License for the specific language governing permissions and +# limitations under the License. + +import collections +import os +import unicodedata +from typing import Iterable, List, Optional, Tuple + +from ...tokenization_utils import PreTrainedTokenizer, _is_control, _is_punctuation, _is_whitespace +from ...utils import logging + + +logger = logging.get_logger(__name__) + +VOCAB_FILES_NAMES = {"vocab_file": "prophetnet.tokenizer"} + + +# Copied from transformers.models.bert.tokenization_bert.whitespace_tokenize +def whitespace_tokenize(text): + """Runs basic whitespace cleaning and splitting on a piece of text.""" + text = text.strip() + if not text: + return [] + tokens = text.split() + return tokens + + +# Copied from transformers.models.bert.tokenization_bert.BasicTokenizer +class BasicTokenizer(object): + """ + Constructs a BasicTokenizer that will run basic tokenization (punctuation splitting, lower casing, etc.). + + Args: + do_lower_case (`bool`, *optional*, defaults to `True`): + Whether or not to lowercase the input when tokenizing. + never_split (`Iterable`, *optional*): + Collection of tokens which will never be split during tokenization. Only has an effect when + `do_basic_tokenize=True` + tokenize_chinese_chars (`bool`, *optional*, defaults to `True`): + Whether or not to tokenize Chinese characters. + + This should likely be deactivated for Japanese (see this + [issue](https://github.com/huggingface/transformers/issues/328)). + strip_accents (`bool`, *optional*): + Whether or not to strip all accents. If this option is not specified, then it will be determined by the + value for `lowercase` (as in the original BERT). + do_split_on_punc (`bool`, *optional*, defaults to `True`): + In some instances we want to skip the basic punctuation splitting so that later tokenization can capture + the full context of the words, such as contractions. + """ + + def __init__( + self, + do_lower_case=True, + never_split=None, + tokenize_chinese_chars=True, + strip_accents=None, + do_split_on_punc=True, + ): + if never_split is None: + never_split = [] + self.do_lower_case = do_lower_case + self.never_split = set(never_split) + self.tokenize_chinese_chars = tokenize_chinese_chars + self.strip_accents = strip_accents + self.do_split_on_punc = do_split_on_punc + + def tokenize(self, text, never_split=None): + """ + Basic Tokenization of a piece of text. For sub-word tokenization, see WordPieceTokenizer. + + Args: + never_split (`List[str]`, *optional*) + Kept for backward compatibility purposes. Now implemented directly at the base class level (see + [`PreTrainedTokenizer.tokenize`]) List of token not to split. + """ + # union() returns a new set by concatenating the two sets. + never_split = self.never_split.union(set(never_split)) if never_split else self.never_split + text = self._clean_text(text) + + # This was added on November 1st, 2018 for the multilingual and Chinese + # models. This is also applied to the English models now, but it doesn't + # matter since the English models were not trained on any Chinese data + # and generally don't have any Chinese data in them (there are Chinese + # characters in the vocabulary because Wikipedia does have some Chinese + # words in the English Wikipedia.). + if self.tokenize_chinese_chars: + text = self._tokenize_chinese_chars(text) + # prevents treating the same character with different unicode codepoints as different characters + unicode_normalized_text = unicodedata.normalize("NFC", text) + orig_tokens = whitespace_tokenize(unicode_normalized_text) + split_tokens = [] + for token in orig_tokens: + if token not in never_split: + if self.do_lower_case: + token = token.lower() + if self.strip_accents is not False: + token = self._run_strip_accents(token) + elif self.strip_accents: + token = self._run_strip_accents(token) + split_tokens.extend(self._run_split_on_punc(token, never_split)) + + output_tokens = whitespace_tokenize(" ".join(split_tokens)) + return output_tokens + + def _run_strip_accents(self, text): + """Strips accents from a piece of text.""" + text = unicodedata.normalize("NFD", text) + output = [] + for char in text: + cat = unicodedata.category(char) + if cat == "Mn": + continue + output.append(char) + return "".join(output) + + def _run_split_on_punc(self, text, never_split=None): + """Splits punctuation on a piece of text.""" + if not self.do_split_on_punc or (never_split is not None and text in never_split): + return [text] + chars = list(text) + i = 0 + start_new_word = True + output = [] + while i < len(chars): + char = chars[i] + if _is_punctuation(char): + output.append([char]) + start_new_word = True + else: + if start_new_word: + output.append([]) + start_new_word = False + output[-1].append(char) + i += 1 + + return ["".join(x) for x in output] + + def _tokenize_chinese_chars(self, text): + """Adds whitespace around any CJK character.""" + output = [] + for char in text: + cp = ord(char) + if self._is_chinese_char(cp): + output.append(" ") + output.append(char) + output.append(" ") + else: + output.append(char) + return "".join(output) + + def _is_chinese_char(self, cp): + """Checks whether CP is the codepoint of a CJK character.""" + # This defines a "chinese character" as anything in the CJK Unicode block: + # https://en.wikipedia.org/wiki/CJK_Unified_Ideographs_(Unicode_block) + # + # Note that the CJK Unicode block is NOT all Japanese and Korean characters, + # despite its name. The modern Korean Hangul alphabet is a different block, + # as is Japanese Hiragana and Katakana. Those alphabets are used to write + # space-separated words, so they are not treated specially and handled + # like the all of the other languages. + if ( + (cp >= 0x4E00 and cp <= 0x9FFF) + or (cp >= 0x3400 and cp <= 0x4DBF) # + or (cp >= 0x20000 and cp <= 0x2A6DF) # + or (cp >= 0x2A700 and cp <= 0x2B73F) # + or (cp >= 0x2B740 and cp <= 0x2B81F) # + or (cp >= 0x2B820 and cp <= 0x2CEAF) # + or (cp >= 0xF900 and cp <= 0xFAFF) + or (cp >= 0x2F800 and cp <= 0x2FA1F) # + ): # + return True + + return False + + def _clean_text(self, text): + """Performs invalid character removal and whitespace cleanup on text.""" + output = [] + for char in text: + cp = ord(char) + if cp == 0 or cp == 0xFFFD or _is_control(char): + continue + if _is_whitespace(char): + output.append(" ") + else: + output.append(char) + return "".join(output) + + +# Copied from transformers.models.bert.tokenization_bert.WordpieceTokenizer +class WordpieceTokenizer(object): + """Runs WordPiece tokenization.""" + + def __init__(self, vocab, unk_token, max_input_chars_per_word=100): + self.vocab = vocab + self.unk_token = unk_token + self.max_input_chars_per_word = max_input_chars_per_word + + def tokenize(self, text): + """ + Tokenizes a piece of text into its word pieces. This uses a greedy longest-match-first algorithm to perform + tokenization using the given vocabulary. + + For example, `input = "unaffable"` wil return as output `["un", "##aff", "##able"]`. + + Args: + text: A single token or whitespace separated tokens. This should have + already been passed through *BasicTokenizer*. + + Returns: + A list of wordpiece tokens. + """ + + output_tokens = [] + for token in whitespace_tokenize(text): + chars = list(token) + if len(chars) > self.max_input_chars_per_word: + output_tokens.append(self.unk_token) + continue + + is_bad = False + start = 0 + sub_tokens = [] + while start < len(chars): + end = len(chars) + cur_substr = None + while start < end: + substr = "".join(chars[start:end]) + if start > 0: + substr = "##" + substr + if substr in self.vocab: + cur_substr = substr + break + end -= 1 + if cur_substr is None: + is_bad = True + break + sub_tokens.append(cur_substr) + start = end + + if is_bad: + output_tokens.append(self.unk_token) + else: + output_tokens.extend(sub_tokens) + return output_tokens + + +def load_vocab(vocab_file): + """Loads a vocabulary file into a dictionary.""" + vocab = collections.OrderedDict() + with open(vocab_file, "r", encoding="utf-8") as reader: + tokens = reader.readlines() + for index, token in enumerate(tokens): + token = token.rstrip("\n") + vocab[token] = index + return vocab + + +class ProphetNetTokenizer(PreTrainedTokenizer): + r""" + Construct a ProphetNetTokenizer. Based on WordPiece. + + This tokenizer inherits from [`PreTrainedTokenizer`] which contains most of the main methods. Users should refer to + this superclass for more information regarding those methods. + + Args: + vocab_file (`str`): + File containing the vocabulary. + do_lower_case (`bool`, *optional*, defaults to `True`): + Whether or not to lowercase the input when tokenizing. + do_basic_tokenize (`bool`, *optional*, defaults to `True`): + Whether or not to do basic tokenization before WordPiece. + never_split (`Iterable`, *optional*): + Collection of tokens which will never be split during tokenization. Only has an effect when + `do_basic_tokenize=True` + unk_token (`str`, *optional*, defaults to `"[UNK]"`): + The unknown token. A token that is not in the vocabulary cannot be converted to an ID and is set to be this + token instead. + sep_token (`str`, *optional*, defaults to `"[SEP]"`): + The separator token, which is used when building a sequence from multiple sequences, e.g. two sequences for + sequence classification or for a text and a question for question answering. It is also used as the last + token of a sequence built with special tokens. + x_sep_token (`str`, *optional*, defaults to `"[X_SEP]"`): + Special second separator token, which can be generated by [`ProphetNetForConditionalGeneration`]. It is + used to separate bullet-point like sentences in summarization, *e.g.*. + pad_token (`str`, *optional*, defaults to `"[PAD]"`): + The token used for padding, for example when batching sequences of different lengths. + mask_token (`str`, *optional*, defaults to `"[MASK]"`): + The token used for masking values. This is the token used when training this model with masked language + modeling. This is the token which the model will try to predict. + tokenize_chinese_chars (`bool`, *optional*, defaults to `True`): + Whether or not to tokenize Chinese characters. + + This should likely be deactivated for Japanese (see this + [issue](https://github.com/huggingface/transformers/issues/328)). + strip_accents (`bool`, *optional*): + Whether or not to strip all accents. If this option is not specified, then it will be determined by the + value for `lowercase` (as in the original BERT). + """ + + vocab_files_names = VOCAB_FILES_NAMES + + # first name has to correspond to main model input name + # to make sure `tokenizer.pad(...)` works correctly + # `ProphetNet` doesn't have `token_type_ids` as argument. + model_input_names: List[str] = ["input_ids", "attention_mask"] + + def __init__( + self, + vocab_file: str, + do_lower_case: Optional[bool] = True, + do_basic_tokenize: Optional[bool] = True, + never_split: Optional[Iterable] = None, + unk_token: Optional[str] = "[UNK]", + sep_token: Optional[str] = "[SEP]", + x_sep_token: Optional[str] = "[X_SEP]", + pad_token: Optional[str] = "[PAD]", + mask_token: Optional[str] = "[MASK]", + tokenize_chinese_chars: Optional[bool] = True, + strip_accents: Optional[bool] = None, + **kwargs, + ): + if not os.path.isfile(vocab_file): + raise ValueError( + f"Can't find a vocabulary file at path '{vocab_file}'. To load the vocabulary from a Google pretrained" + " model use `tokenizer = AutoTokenizer.from_pretrained(PRETRAINED_MODEL_NAME)`" + ) + self.vocab = load_vocab(vocab_file) + self.ids_to_tokens = collections.OrderedDict([(ids, tok) for tok, ids in self.vocab.items()]) + self.do_basic_tokenize = do_basic_tokenize + if do_basic_tokenize: + self.basic_tokenizer = BasicTokenizer( + do_lower_case=do_lower_case, + never_split=never_split, + tokenize_chinese_chars=tokenize_chinese_chars, + strip_accents=strip_accents, + ) + self.wordpiece_tokenizer = WordpieceTokenizer(vocab=self.vocab, unk_token=str(unk_token)) + + super().__init__( + do_lower_case=do_lower_case, + do_basic_tokenize=do_basic_tokenize, + never_split=never_split, + unk_token=unk_token, + sep_token=sep_token, + x_sep_token=x_sep_token, + pad_token=pad_token, + mask_token=mask_token, + tokenize_chinese_chars=tokenize_chinese_chars, + strip_accents=strip_accents, + **kwargs, + ) + + @property + def vocab_size(self): + return len(self.vocab) + + def get_vocab(self): + return dict(self.vocab, **self.added_tokens_encoder) + + def _tokenize(self, text): + split_tokens = [] + if self.do_basic_tokenize: + for token in self.basic_tokenizer.tokenize(text, never_split=self.all_special_tokens): + # If the token is part of the never_split set + if token in self.basic_tokenizer.never_split: + split_tokens.append(token) + else: + split_tokens += self.wordpiece_tokenizer.tokenize(token) + else: + split_tokens = self.wordpiece_tokenizer.tokenize(text) + return split_tokens + + def _convert_token_to_id(self, token: str): + """Converts a token (str) in an id using the vocab.""" + return self.vocab.get(token, self.vocab.get(self.unk_token)) + + def _convert_id_to_token(self, index: int): + """Converts an index (integer) in a token (str) using the vocab.""" + return self.ids_to_tokens.get(index, self.unk_token) + + def convert_tokens_to_string(self, tokens: str): + """Converts a sequence of tokens (string) in a single string.""" + out_string = " ".join(tokens).replace(" ##", "").strip() + return out_string + + def get_special_tokens_mask( + self, + token_ids_0: List[int], + token_ids_1: Optional[List[int]] = None, + already_has_special_tokens: Optional[bool] = False, + ) -> List[int]: + """ + Retrieve sequence ids from a token list that has no special tokens added. This method is called when adding + special tokens using the tokenizer `prepare_for_model` method. + + Args: + token_ids_0 (`List[int]`): + List of IDs. + token_ids_1 (`List[int]`, *optional*): + Optional second list of IDs for sequence pairs. + already_has_special_tokens (`bool`, *optional*, defaults to `False`): + Whether or not the token list is already formatted with special tokens for the model. + + Returns: + `List[int]`: A list of integers in the range [0, 1]: 1 for a special token, 0 for a sequence token. + """ + if already_has_special_tokens: + return super().get_special_tokens_mask( + token_ids_0=token_ids_0, token_ids_1=token_ids_1, already_has_special_tokens=True + ) + + if token_ids_1 is None: + return ([0] * len(token_ids_0)) + [1] + return ([0] * len(token_ids_0)) + [1] + ([0] * len(token_ids_1)) + [1] + + def create_token_type_ids_from_sequences( + self, token_ids_0: List[int], token_ids_1: Optional[List[int]] = None + ) -> List[int]: + """ + Create a mask from the two sequences passed to be used in a sequence-pair classification task. A ProphetNet + sequence pair mask has the following format: + + ``` + 0 0 0 0 0 0 0 0 0 0 0 1 1 1 1 1 1 1 1 1 + | first sequence | second sequence | + ``` + + If `token_ids_1` is `None`, this method only returns the first portion of the mask (0s). + + Args: + token_ids_0 (`List[int]`): + List of IDs. + token_ids_1 (`List[int]`, *optional*): + Optional second list of IDs for sequence pairs. + + Returns: + `List[int]`: List of [token type IDs](../glossary#token-type-ids) according to the given sequence(s). + """ + sep = [self.sep_token_id] + if token_ids_1 is None: + return len(token_ids_0 + sep) * [0] + return len(token_ids_0 + sep) * [0] + len(token_ids_1 + sep) * [1] + + def save_vocabulary(self, save_directory: str, filename_prefix: Optional[str] = None) -> Tuple[str]: + index = 0 + if os.path.isdir(save_directory): + vocab_file = os.path.join( + save_directory, (filename_prefix + "-" if filename_prefix else "") + VOCAB_FILES_NAMES["vocab_file"] + ) + else: + vocab_file = (filename_prefix + "-" if filename_prefix else "") + save_directory + with open(vocab_file, "w", encoding="utf-8") as writer: + for token, token_index in sorted(self.vocab.items(), key=lambda kv: kv[1]): + if index != token_index: + logger.warning( + f"Saving vocabulary to {vocab_file}: vocabulary indices are not consecutive." + " Please check that the vocabulary is not corrupted!" + ) + index = token_index + writer.write(token + "\n") + index += 1 + return (vocab_file,) + + def build_inputs_with_special_tokens( + self, token_ids_0: List[int], token_ids_1: Optional[List[int]] = None + ) -> List[int]: + """ + Build model inputs from a sequence or a pair of sequence for sequence classification tasks by concatenating and + adding special tokens. A BERT sequence has the following format: + + - single sequence: `[CLS] X [SEP]` + - pair of sequences: `[CLS] A [SEP] B [SEP]` + + Args: + token_ids_0 (`List[int]`): + List of IDs to which the special tokens will be added. + token_ids_1 (`List[int]`, *optional*): + Optional second list of IDs for sequence pairs. + + Returns: + `List[int]`: List of [input IDs](../glossary#input-ids) with the appropriate special tokens. + """ + if token_ids_1 is None: + return token_ids_0 + [self.sep_token_id] + sep = [self.sep_token_id] + return token_ids_0 + sep + token_ids_1 + sep diff --git a/venv/lib/python3.10/site-packages/transformers/models/upernet/__init__.py b/venv/lib/python3.10/site-packages/transformers/models/upernet/__init__.py new file mode 100644 index 0000000000000000000000000000000000000000..3954fe4594dad04c3908a447f36dd02a1dea8c7c --- /dev/null +++ b/venv/lib/python3.10/site-packages/transformers/models/upernet/__init__.py @@ -0,0 +1,50 @@ +# Copyright 2022 The HuggingFace Team. All rights reserved. +# +# Licensed under the Apache License, Version 2.0 (the "License"); +# you may not use this file except in compliance with the License. +# You may obtain a copy of the License at +# +# http://www.apache.org/licenses/LICENSE-2.0 +# +# Unless required by applicable law or agreed to in writing, software +# distributed under the License is distributed on an "AS IS" BASIS, +# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. +# See the License for the specific language governing permissions and +# limitations under the License. +from typing import TYPE_CHECKING + +from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_torch_available + + +_import_structure = { + "configuration_upernet": ["UperNetConfig"], +} + +try: + if not is_torch_available(): + raise OptionalDependencyNotAvailable() +except OptionalDependencyNotAvailable: + pass +else: + _import_structure["modeling_upernet"] = [ + "UperNetForSemanticSegmentation", + "UperNetPreTrainedModel", + ] + + +if TYPE_CHECKING: + from .configuration_upernet import UperNetConfig + + try: + if not is_torch_available(): + raise OptionalDependencyNotAvailable() + except OptionalDependencyNotAvailable: + pass + else: + from .modeling_upernet import UperNetForSemanticSegmentation, UperNetPreTrainedModel + + +else: + import sys + + sys.modules[__name__] = _LazyModule(__name__, globals()["__file__"], _import_structure, module_spec=__spec__) diff --git a/venv/lib/python3.10/site-packages/transformers/models/upernet/__pycache__/__init__.cpython-310.pyc b/venv/lib/python3.10/site-packages/transformers/models/upernet/__pycache__/__init__.cpython-310.pyc new file mode 100644 index 0000000000000000000000000000000000000000..fd828ad6880c805f58024d7258c70b57d7f5e8f2 Binary files /dev/null and b/venv/lib/python3.10/site-packages/transformers/models/upernet/__pycache__/__init__.cpython-310.pyc differ diff --git a/venv/lib/python3.10/site-packages/transformers/models/upernet/__pycache__/configuration_upernet.cpython-310.pyc b/venv/lib/python3.10/site-packages/transformers/models/upernet/__pycache__/configuration_upernet.cpython-310.pyc new file mode 100644 index 0000000000000000000000000000000000000000..d6373689a0bcbdcf7077eb87e4f4ceda1e3d22c0 Binary files /dev/null and b/venv/lib/python3.10/site-packages/transformers/models/upernet/__pycache__/configuration_upernet.cpython-310.pyc differ diff --git a/venv/lib/python3.10/site-packages/transformers/models/upernet/__pycache__/convert_convnext_upernet_to_pytorch.cpython-310.pyc b/venv/lib/python3.10/site-packages/transformers/models/upernet/__pycache__/convert_convnext_upernet_to_pytorch.cpython-310.pyc new file mode 100644 index 0000000000000000000000000000000000000000..270b6ae766a0f1a6c103033bb98571c60fafe928 Binary files /dev/null and b/venv/lib/python3.10/site-packages/transformers/models/upernet/__pycache__/convert_convnext_upernet_to_pytorch.cpython-310.pyc differ diff --git a/venv/lib/python3.10/site-packages/transformers/models/upernet/__pycache__/convert_swin_upernet_to_pytorch.cpython-310.pyc b/venv/lib/python3.10/site-packages/transformers/models/upernet/__pycache__/convert_swin_upernet_to_pytorch.cpython-310.pyc new file mode 100644 index 0000000000000000000000000000000000000000..7ce4606249a02dad6ee9ecb78455e512a4fb986d Binary files /dev/null and b/venv/lib/python3.10/site-packages/transformers/models/upernet/__pycache__/convert_swin_upernet_to_pytorch.cpython-310.pyc differ diff --git a/venv/lib/python3.10/site-packages/transformers/models/upernet/__pycache__/modeling_upernet.cpython-310.pyc b/venv/lib/python3.10/site-packages/transformers/models/upernet/__pycache__/modeling_upernet.cpython-310.pyc new file mode 100644 index 0000000000000000000000000000000000000000..e6f5b98fc733187159d1ce010b4f033b82fcdf13 Binary files /dev/null and b/venv/lib/python3.10/site-packages/transformers/models/upernet/__pycache__/modeling_upernet.cpython-310.pyc differ diff --git a/venv/lib/python3.10/site-packages/transformers/models/upernet/configuration_upernet.py b/venv/lib/python3.10/site-packages/transformers/models/upernet/configuration_upernet.py new file mode 100644 index 0000000000000000000000000000000000000000..609818c80d17b728322d52586ab1cfedf687f8e0 --- /dev/null +++ b/venv/lib/python3.10/site-packages/transformers/models/upernet/configuration_upernet.py @@ -0,0 +1,138 @@ +# coding=utf-8 +# Copyright 2022 The HuggingFace Inc. team. All rights reserved. +# +# Licensed under the Apache License, Version 2.0 (the "License"); +# you may not use this file except in compliance with the License. +# You may obtain a copy of the License at +# +# http://www.apache.org/licenses/LICENSE-2.0 +# +# Unless required by applicable law or agreed to in writing, software +# distributed under the License is distributed on an "AS IS" BASIS, +# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. +# See the License for the specific language governing permissions and +# limitations under the License. +""" UperNet model configuration""" + + +from ...configuration_utils import PretrainedConfig +from ...utils import logging +from ..auto.configuration_auto import CONFIG_MAPPING + + +logger = logging.get_logger(__name__) + + +class UperNetConfig(PretrainedConfig): + r""" + This is the configuration class to store the configuration of an [`UperNetForSemanticSegmentation`]. It is used to + instantiate an UperNet model according to the specified arguments, defining the model architecture. Instantiating a + configuration with the defaults will yield a similar configuration to that of the UperNet + [openmmlab/upernet-convnext-tiny](https://huggingface.co/openmmlab/upernet-convnext-tiny) architecture. + + Configuration objects inherit from [`PretrainedConfig`] and can be used to control the model outputs. Read the + documentation from [`PretrainedConfig`] for more information. + + Args: + backbone_config (`PretrainedConfig` or `dict`, *optional*, defaults to `ResNetConfig()`): + The configuration of the backbone model. + backbone (`str`, *optional*): + Name of backbone to use when `backbone_config` is `None`. If `use_pretrained_backbone` is `True`, this + will load the corresponding pretrained weights from the timm or transformers library. If `use_pretrained_backbone` + is `False`, this loads the backbone's config and uses that to initialize the backbone with random weights. + use_pretrained_backbone (`bool`, *optional*, `False`): + Whether to use pretrained weights for the backbone. + use_timm_backbone (`bool`, *optional*, `False`): + Whether to load `backbone` from the timm library. If `False`, the backbone is loaded from the transformers + library. + backbone_kwargs (`dict`, *optional*): + Keyword arguments to be passed to AutoBackbone when loading from a checkpoint + e.g. `{'out_indices': (0, 1, 2, 3)}`. Cannot be specified if `backbone_config` is set. + hidden_size (`int`, *optional*, defaults to 512): + The number of hidden units in the convolutional layers. + initializer_range (`float`, *optional*, defaults to 0.02): + The standard deviation of the truncated_normal_initializer for initializing all weight matrices. + pool_scales (`Tuple[int]`, *optional*, defaults to `[1, 2, 3, 6]`): + Pooling scales used in Pooling Pyramid Module applied on the last feature map. + use_auxiliary_head (`bool`, *optional*, defaults to `True`): + Whether to use an auxiliary head during training. + auxiliary_loss_weight (`float`, *optional*, defaults to 0.4): + Weight of the cross-entropy loss of the auxiliary head. + auxiliary_channels (`int`, *optional*, defaults to 256): + Number of channels to use in the auxiliary head. + auxiliary_num_convs (`int`, *optional*, defaults to 1): + Number of convolutional layers to use in the auxiliary head. + auxiliary_concat_input (`bool`, *optional*, defaults to `False`): + Whether to concatenate the output of the auxiliary head with the input before the classification layer. + loss_ignore_index (`int`, *optional*, defaults to 255): + The index that is ignored by the loss function. + + Examples: + + ```python + >>> from transformers import UperNetConfig, UperNetForSemanticSegmentation + + >>> # Initializing a configuration + >>> configuration = UperNetConfig() + + >>> # Initializing a model (with random weights) from the configuration + >>> model = UperNetForSemanticSegmentation(configuration) + + >>> # Accessing the model configuration + >>> configuration = model.config + ```""" + + model_type = "upernet" + + def __init__( + self, + backbone_config=None, + backbone=None, + use_pretrained_backbone=False, + use_timm_backbone=False, + backbone_kwargs=None, + hidden_size=512, + initializer_range=0.02, + pool_scales=[1, 2, 3, 6], + use_auxiliary_head=True, + auxiliary_loss_weight=0.4, + auxiliary_in_channels=384, + auxiliary_channels=256, + auxiliary_num_convs=1, + auxiliary_concat_input=False, + loss_ignore_index=255, + **kwargs, + ): + super().__init__(**kwargs) + if use_pretrained_backbone: + raise ValueError("Pretrained backbones are not supported yet.") + + if backbone_config is not None and backbone is not None: + raise ValueError("You can't specify both `backbone` and `backbone_config`.") + + if backbone_config is None and backbone is None: + logger.info("`backbone_config` is `None`. Initializing the config with the default `ResNet` backbone.") + backbone_config = CONFIG_MAPPING["resnet"](out_features=["stage1", "stage2", "stage3", "stage4"]) + elif isinstance(backbone_config, dict): + backbone_model_type = backbone_config.get("model_type") + config_class = CONFIG_MAPPING[backbone_model_type] + backbone_config = config_class.from_dict(backbone_config) + + if backbone_kwargs is not None and backbone_kwargs and backbone_config is not None: + raise ValueError("You can't specify both `backbone_kwargs` and `backbone_config`.") + + self.backbone_config = backbone_config + self.backbone = backbone + self.use_pretrained_backbone = use_pretrained_backbone + self.use_timm_backbone = use_timm_backbone + self.backbone_kwargs = backbone_kwargs + self.hidden_size = hidden_size + self.initializer_range = initializer_range + self.pool_scales = pool_scales + self.use_auxiliary_head = use_auxiliary_head + self.auxiliary_loss_weight = auxiliary_loss_weight + self.auxiliary_in_channels = auxiliary_in_channels + self.auxiliary_channels = auxiliary_channels + self.auxiliary_num_convs = auxiliary_num_convs + self.auxiliary_concat_input = auxiliary_concat_input + self.loss_ignore_index = loss_ignore_index diff --git a/venv/lib/python3.10/site-packages/transformers/models/upernet/convert_convnext_upernet_to_pytorch.py b/venv/lib/python3.10/site-packages/transformers/models/upernet/convert_convnext_upernet_to_pytorch.py new file mode 100644 index 0000000000000000000000000000000000000000..eeb3ab5fc9938171099a47feef23c4694d8b5169 --- /dev/null +++ b/venv/lib/python3.10/site-packages/transformers/models/upernet/convert_convnext_upernet_to_pytorch.py @@ -0,0 +1,214 @@ +# coding=utf-8 +# Copyright 2022 The HuggingFace Inc. team. +# +# Licensed under the Apache License, Version 2.0 (the "License"); +# you may not use this file except in compliance with the License. +# You may obtain a copy of the License at +# +# http://www.apache.org/licenses/LICENSE-2.0 +# +# Unless required by applicable law or agreed to in writing, software +# distributed under the License is distributed on an "AS IS" BASIS, +# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. +# See the License for the specific language governing permissions and +# limitations under the License. +"""Convert ConvNext + UperNet checkpoints from mmsegmentation.""" + +import argparse +import json + +import requests +import torch +from huggingface_hub import hf_hub_download +from PIL import Image + +from transformers import ConvNextConfig, SegformerImageProcessor, UperNetConfig, UperNetForSemanticSegmentation + + +def get_upernet_config(model_name): + auxiliary_in_channels = 384 + if "tiny" in model_name: + depths = [3, 3, 9, 3] + hidden_sizes = [96, 192, 384, 768] + if "small" in model_name: + depths = [3, 3, 27, 3] + hidden_sizes = [96, 192, 384, 768] + if "base" in model_name: + depths = [3, 3, 27, 3] + hidden_sizes = [128, 256, 512, 1024] + auxiliary_in_channels = 512 + if "large" in model_name: + depths = [3, 3, 27, 3] + hidden_sizes = [192, 384, 768, 1536] + auxiliary_in_channels = 768 + if "xlarge" in model_name: + depths = [3, 3, 27, 3] + hidden_sizes = [256, 512, 1024, 2048] + auxiliary_in_channels = 1024 + + # set label information + num_labels = 150 + repo_id = "huggingface/label-files" + filename = "ade20k-id2label.json" + id2label = json.load(open(hf_hub_download(repo_id, filename, repo_type="dataset"), "r")) + id2label = {int(k): v for k, v in id2label.items()} + label2id = {v: k for k, v in id2label.items()} + + backbone_config = ConvNextConfig( + depths=depths, hidden_sizes=hidden_sizes, out_features=["stage1", "stage2", "stage3", "stage4"] + ) + config = UperNetConfig( + backbone_config=backbone_config, + auxiliary_in_channels=auxiliary_in_channels, + num_labels=num_labels, + id2label=id2label, + label2id=label2id, + ) + + return config + + +# here we list all keys to be renamed (original name on the left, our name on the right) +def create_rename_keys(config): + rename_keys = [] + + # fmt: off + # stem + rename_keys.append(("backbone.downsample_layers.0.0.weight", "backbone.embeddings.patch_embeddings.weight")) + rename_keys.append(("backbone.downsample_layers.0.0.bias", "backbone.embeddings.patch_embeddings.bias")) + rename_keys.append(("backbone.downsample_layers.0.1.weight", "backbone.embeddings.layernorm.weight")) + rename_keys.append(("backbone.downsample_layers.0.1.bias", "backbone.embeddings.layernorm.bias")) + # stages + for i in range(len(config.backbone_config.depths)): + for j in range(config.backbone_config.depths[i]): + rename_keys.append((f"backbone.stages.{i}.{j}.gamma", f"backbone.encoder.stages.{i}.layers.{j}.layer_scale_parameter")) + rename_keys.append((f"backbone.stages.{i}.{j}.depthwise_conv.weight", f"backbone.encoder.stages.{i}.layers.{j}.dwconv.weight")) + rename_keys.append((f"backbone.stages.{i}.{j}.depthwise_conv.bias", f"backbone.encoder.stages.{i}.layers.{j}.dwconv.bias")) + rename_keys.append((f"backbone.stages.{i}.{j}.norm.weight", f"backbone.encoder.stages.{i}.layers.{j}.layernorm.weight")) + rename_keys.append((f"backbone.stages.{i}.{j}.norm.bias", f"backbone.encoder.stages.{i}.layers.{j}.layernorm.bias")) + rename_keys.append((f"backbone.stages.{i}.{j}.pointwise_conv1.weight", f"backbone.encoder.stages.{i}.layers.{j}.pwconv1.weight")) + rename_keys.append((f"backbone.stages.{i}.{j}.pointwise_conv1.bias", f"backbone.encoder.stages.{i}.layers.{j}.pwconv1.bias")) + rename_keys.append((f"backbone.stages.{i}.{j}.pointwise_conv2.weight", f"backbone.encoder.stages.{i}.layers.{j}.pwconv2.weight")) + rename_keys.append((f"backbone.stages.{i}.{j}.pointwise_conv2.bias", f"backbone.encoder.stages.{i}.layers.{j}.pwconv2.bias")) + if i > 0: + rename_keys.append((f"backbone.downsample_layers.{i}.0.weight", f"backbone.encoder.stages.{i}.downsampling_layer.0.weight")) + rename_keys.append((f"backbone.downsample_layers.{i}.0.bias", f"backbone.encoder.stages.{i}.downsampling_layer.0.bias")) + rename_keys.append((f"backbone.downsample_layers.{i}.1.weight", f"backbone.encoder.stages.{i}.downsampling_layer.1.weight")) + rename_keys.append((f"backbone.downsample_layers.{i}.1.bias", f"backbone.encoder.stages.{i}.downsampling_layer.1.bias")) + + rename_keys.append((f"backbone.norm{i}.weight", f"backbone.hidden_states_norms.stage{i+1}.weight")) + rename_keys.append((f"backbone.norm{i}.bias", f"backbone.hidden_states_norms.stage{i+1}.bias")) + + # decode head + rename_keys.extend( + [ + ("decode_head.conv_seg.weight", "decode_head.classifier.weight"), + ("decode_head.conv_seg.bias", "decode_head.classifier.bias"), + ("auxiliary_head.conv_seg.weight", "auxiliary_head.classifier.weight"), + ("auxiliary_head.conv_seg.bias", "auxiliary_head.classifier.bias"), + ] + ) + # fmt: on + + return rename_keys + + +def rename_key(dct, old, new): + val = dct.pop(old) + dct[new] = val + + +def convert_upernet_checkpoint(model_name, pytorch_dump_folder_path, push_to_hub): + model_name_to_url = { + "upernet-convnext-tiny": "https://download.openmmlab.com/mmsegmentation/v0.5/convnext/upernet_convnext_tiny_fp16_512x512_160k_ade20k/upernet_convnext_tiny_fp16_512x512_160k_ade20k_20220227_124553-cad485de.pth", + "upernet-convnext-small": "https://download.openmmlab.com/mmsegmentation/v0.5/convnext/upernet_convnext_small_fp16_512x512_160k_ade20k/upernet_convnext_small_fp16_512x512_160k_ade20k_20220227_131208-1b1e394f.pth", + "upernet-convnext-base": "https://download.openmmlab.com/mmsegmentation/v0.5/convnext/upernet_convnext_base_fp16_512x512_160k_ade20k/upernet_convnext_base_fp16_512x512_160k_ade20k_20220227_181227-02a24fc6.pth", + "upernet-convnext-large": "https://download.openmmlab.com/mmsegmentation/v0.5/convnext/upernet_convnext_large_fp16_640x640_160k_ade20k/upernet_convnext_large_fp16_640x640_160k_ade20k_20220226_040532-e57aa54d.pth", + "upernet-convnext-xlarge": "https://download.openmmlab.com/mmsegmentation/v0.5/convnext/upernet_convnext_xlarge_fp16_640x640_160k_ade20k/upernet_convnext_xlarge_fp16_640x640_160k_ade20k_20220226_080344-95fc38c2.pth", + } + checkpoint_url = model_name_to_url[model_name] + state_dict = torch.hub.load_state_dict_from_url(checkpoint_url, map_location="cpu")["state_dict"] + + config = get_upernet_config(model_name) + model = UperNetForSemanticSegmentation(config) + model.eval() + + # replace "bn" => "batch_norm" + for key in state_dict.copy().keys(): + val = state_dict.pop(key) + if "bn" in key: + key = key.replace("bn", "batch_norm") + state_dict[key] = val + + # rename keys + rename_keys = create_rename_keys(config) + for src, dest in rename_keys: + rename_key(state_dict, src, dest) + + model.load_state_dict(state_dict) + + # verify on image + url = "https://huggingface.co/datasets/hf-internal-testing/fixtures_ade20k/resolve/main/ADE_val_00000001.jpg" + image = Image.open(requests.get(url, stream=True).raw).convert("RGB") + + processor = SegformerImageProcessor() + pixel_values = processor(image, return_tensors="pt").pixel_values + + with torch.no_grad(): + outputs = model(pixel_values) + + if model_name == "upernet-convnext-tiny": + expected_slice = torch.tensor( + [[-8.8110, -8.8110, -8.6521], [-8.8110, -8.8110, -8.6521], [-8.7746, -8.7746, -8.6130]] + ) + elif model_name == "upernet-convnext-small": + expected_slice = torch.tensor( + [[-8.8236, -8.8236, -8.6771], [-8.8236, -8.8236, -8.6771], [-8.7638, -8.7638, -8.6240]] + ) + elif model_name == "upernet-convnext-base": + expected_slice = torch.tensor( + [[-8.8558, -8.8558, -8.6905], [-8.8558, -8.8558, -8.6905], [-8.7669, -8.7669, -8.6021]] + ) + elif model_name == "upernet-convnext-large": + expected_slice = torch.tensor( + [[-8.6660, -8.6660, -8.6210], [-8.6660, -8.6660, -8.6210], [-8.6310, -8.6310, -8.5964]] + ) + elif model_name == "upernet-convnext-xlarge": + expected_slice = torch.tensor( + [[-8.4980, -8.4980, -8.3977], [-8.4980, -8.4980, -8.3977], [-8.4379, -8.4379, -8.3412]] + ) + print("Logits:", outputs.logits[0, 0, :3, :3]) + assert torch.allclose(outputs.logits[0, 0, :3, :3], expected_slice, atol=1e-4) + print("Looks ok!") + + if pytorch_dump_folder_path is not None: + print(f"Saving model {model_name} to {pytorch_dump_folder_path}") + model.save_pretrained(pytorch_dump_folder_path) + print(f"Saving processor to {pytorch_dump_folder_path}") + processor.save_pretrained(pytorch_dump_folder_path) + + if push_to_hub: + print(f"Pushing model and processor for {model_name} to hub") + model.push_to_hub(f"openmmlab/{model_name}") + processor.push_to_hub(f"openmmlab/{model_name}") + + +if __name__ == "__main__": + parser = argparse.ArgumentParser() + # Required parameters + parser.add_argument( + "--model_name", + default="upernet-convnext-tiny", + type=str, + choices=[f"upernet-convnext-{size}" for size in ["tiny", "small", "base", "large", "xlarge"]], + help="Name of the ConvNext UperNet model you'd like to convert.", + ) + parser.add_argument( + "--pytorch_dump_folder_path", default=None, type=str, help="Path to the output PyTorch model directory." + ) + parser.add_argument( + "--push_to_hub", action="store_true", help="Whether or not to push the converted model to the 🤗 hub." + ) + + args = parser.parse_args() + convert_upernet_checkpoint(args.model_name, args.pytorch_dump_folder_path, args.push_to_hub) diff --git a/venv/lib/python3.10/site-packages/transformers/models/upernet/convert_swin_upernet_to_pytorch.py b/venv/lib/python3.10/site-packages/transformers/models/upernet/convert_swin_upernet_to_pytorch.py new file mode 100644 index 0000000000000000000000000000000000000000..9580af7c46a50c26c25fe5a9f2728188fbd0193e --- /dev/null +++ b/venv/lib/python3.10/site-packages/transformers/models/upernet/convert_swin_upernet_to_pytorch.py @@ -0,0 +1,297 @@ +# coding=utf-8 +# Copyright 2022 The HuggingFace Inc. team. +# +# Licensed under the Apache License, Version 2.0 (the "License"); +# you may not use this file except in compliance with the License. +# You may obtain a copy of the License at +# +# http://www.apache.org/licenses/LICENSE-2.0 +# +# Unless required by applicable law or agreed to in writing, software +# distributed under the License is distributed on an "AS IS" BASIS, +# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. +# See the License for the specific language governing permissions and +# limitations under the License. +"""Convert Swin Transformer + UperNet checkpoints from mmsegmentation. + +URL: https://github.com/open-mmlab/mmsegmentation/tree/master/configs/swin +""" + +import argparse +import json + +import requests +import torch +from huggingface_hub import hf_hub_download +from PIL import Image + +from transformers import SegformerImageProcessor, SwinConfig, UperNetConfig, UperNetForSemanticSegmentation + + +def get_upernet_config(model_name): + auxiliary_in_channels = 384 + window_size = 7 + if "tiny" in model_name: + embed_dim = 96 + depths = (2, 2, 6, 2) + num_heads = (3, 6, 12, 24) + elif "small" in model_name: + embed_dim = 96 + depths = (2, 2, 18, 2) + num_heads = (3, 6, 12, 24) + elif "base" in model_name: + embed_dim = 128 + depths = (2, 2, 18, 2) + num_heads = (4, 8, 16, 32) + window_size = 12 + auxiliary_in_channels = 512 + elif "large" in model_name: + embed_dim = 192 + depths = (2, 2, 18, 2) + num_heads = (6, 12, 24, 48) + window_size = 12 + auxiliary_in_channels = 768 + + # set label information + num_labels = 150 + repo_id = "huggingface/label-files" + filename = "ade20k-id2label.json" + id2label = json.load(open(hf_hub_download(repo_id, filename, repo_type="dataset"), "r")) + id2label = {int(k): v for k, v in id2label.items()} + label2id = {v: k for k, v in id2label.items()} + + backbone_config = SwinConfig( + embed_dim=embed_dim, + depths=depths, + num_heads=num_heads, + window_size=window_size, + out_features=["stage1", "stage2", "stage3", "stage4"], + ) + config = UperNetConfig( + backbone_config=backbone_config, + auxiliary_in_channels=auxiliary_in_channels, + num_labels=num_labels, + id2label=id2label, + label2id=label2id, + ) + + return config + + +# here we list all keys to be renamed (original name on the left, our name on the right) +def create_rename_keys(config): + rename_keys = [] + + # fmt: off + # stem + rename_keys.append(("backbone.patch_embed.projection.weight", "backbone.embeddings.patch_embeddings.projection.weight")) + rename_keys.append(("backbone.patch_embed.projection.bias", "backbone.embeddings.patch_embeddings.projection.bias")) + rename_keys.append(("backbone.patch_embed.norm.weight", "backbone.embeddings.norm.weight")) + rename_keys.append(("backbone.patch_embed.norm.bias", "backbone.embeddings.norm.bias")) + # stages + for i in range(len(config.backbone_config.depths)): + for j in range(config.backbone_config.depths[i]): + rename_keys.append((f"backbone.stages.{i}.blocks.{j}.norm1.weight", f"backbone.encoder.layers.{i}.blocks.{j}.layernorm_before.weight")) + rename_keys.append((f"backbone.stages.{i}.blocks.{j}.norm1.bias", f"backbone.encoder.layers.{i}.blocks.{j}.layernorm_before.bias")) + rename_keys.append((f"backbone.stages.{i}.blocks.{j}.attn.w_msa.relative_position_bias_table", f"backbone.encoder.layers.{i}.blocks.{j}.attention.self.relative_position_bias_table")) + rename_keys.append((f"backbone.stages.{i}.blocks.{j}.attn.w_msa.relative_position_index", f"backbone.encoder.layers.{i}.blocks.{j}.attention.self.relative_position_index")) + rename_keys.append((f"backbone.stages.{i}.blocks.{j}.attn.w_msa.proj.weight", f"backbone.encoder.layers.{i}.blocks.{j}.attention.output.dense.weight")) + rename_keys.append((f"backbone.stages.{i}.blocks.{j}.attn.w_msa.proj.bias", f"backbone.encoder.layers.{i}.blocks.{j}.attention.output.dense.bias")) + rename_keys.append((f"backbone.stages.{i}.blocks.{j}.norm2.weight", f"backbone.encoder.layers.{i}.blocks.{j}.layernorm_after.weight")) + rename_keys.append((f"backbone.stages.{i}.blocks.{j}.norm2.bias", f"backbone.encoder.layers.{i}.blocks.{j}.layernorm_after.bias")) + rename_keys.append((f"backbone.stages.{i}.blocks.{j}.ffn.layers.0.0.weight", f"backbone.encoder.layers.{i}.blocks.{j}.intermediate.dense.weight")) + rename_keys.append((f"backbone.stages.{i}.blocks.{j}.ffn.layers.0.0.bias", f"backbone.encoder.layers.{i}.blocks.{j}.intermediate.dense.bias")) + rename_keys.append((f"backbone.stages.{i}.blocks.{j}.ffn.layers.1.weight", f"backbone.encoder.layers.{i}.blocks.{j}.output.dense.weight")) + rename_keys.append((f"backbone.stages.{i}.blocks.{j}.ffn.layers.1.bias", f"backbone.encoder.layers.{i}.blocks.{j}.output.dense.bias")) + + if i < 3: + rename_keys.append((f"backbone.stages.{i}.downsample.reduction.weight", f"backbone.encoder.layers.{i}.downsample.reduction.weight")) + rename_keys.append((f"backbone.stages.{i}.downsample.norm.weight", f"backbone.encoder.layers.{i}.downsample.norm.weight")) + rename_keys.append((f"backbone.stages.{i}.downsample.norm.bias", f"backbone.encoder.layers.{i}.downsample.norm.bias")) + rename_keys.append((f"backbone.norm{i}.weight", f"backbone.hidden_states_norms.stage{i+1}.weight")) + rename_keys.append((f"backbone.norm{i}.bias", f"backbone.hidden_states_norms.stage{i+1}.bias")) + + # decode head + rename_keys.extend( + [ + ("decode_head.conv_seg.weight", "decode_head.classifier.weight"), + ("decode_head.conv_seg.bias", "decode_head.classifier.bias"), + ("auxiliary_head.conv_seg.weight", "auxiliary_head.classifier.weight"), + ("auxiliary_head.conv_seg.bias", "auxiliary_head.classifier.bias"), + ] + ) + # fmt: on + + return rename_keys + + +def rename_key(dct, old, new): + val = dct.pop(old) + dct[new] = val + + +# we split up the matrix of each encoder layer into queries, keys and values +def read_in_q_k_v(state_dict, backbone_config): + num_features = [int(backbone_config.embed_dim * 2**i) for i in range(len(backbone_config.depths))] + for i in range(len(backbone_config.depths)): + dim = num_features[i] + for j in range(backbone_config.depths[i]): + # fmt: off + # read in weights + bias of input projection layer (in original implementation, this is a single matrix + bias) + in_proj_weight = state_dict.pop(f"backbone.stages.{i}.blocks.{j}.attn.w_msa.qkv.weight") + in_proj_bias = state_dict.pop(f"backbone.stages.{i}.blocks.{j}.attn.w_msa.qkv.bias") + # next, add query, keys and values (in that order) to the state dict + state_dict[f"backbone.encoder.layers.{i}.blocks.{j}.attention.self.query.weight"] = in_proj_weight[:dim, :] + state_dict[f"backbone.encoder.layers.{i}.blocks.{j}.attention.self.query.bias"] = in_proj_bias[: dim] + state_dict[f"backbone.encoder.layers.{i}.blocks.{j}.attention.self.key.weight"] = in_proj_weight[ + dim : dim * 2, : + ] + state_dict[f"backbone.encoder.layers.{i}.blocks.{j}.attention.self.key.bias"] = in_proj_bias[ + dim : dim * 2 + ] + state_dict[f"backbone.encoder.layers.{i}.blocks.{j}.attention.self.value.weight"] = in_proj_weight[ + -dim :, : + ] + state_dict[f"backbone.encoder.layers.{i}.blocks.{j}.attention.self.value.bias"] = in_proj_bias[-dim :] + # fmt: on + + +def correct_unfold_reduction_order(x): + out_channel, in_channel = x.shape + x = x.reshape(out_channel, 4, in_channel // 4) + x = x[:, [0, 2, 1, 3], :].transpose(1, 2).reshape(out_channel, in_channel) + return x + + +def reverse_correct_unfold_reduction_order(x): + out_channel, in_channel = x.shape + x = x.reshape(out_channel, in_channel // 4, 4) + x = x[:, :, [0, 2, 1, 3]].transpose(1, 2).reshape(out_channel, in_channel) + + return x + + +def correct_unfold_norm_order(x): + in_channel = x.shape[0] + x = x.reshape(4, in_channel // 4) + x = x[[0, 2, 1, 3], :].transpose(0, 1).reshape(in_channel) + return x + + +# there was an incompatibility with this version, due to a new implementation of their downsampling operation using nn.Unfold. +# was resolved as seen here: +# https://github.com/open-mmlab/mmdetection/blob/31c84958f54287a8be2b99cbf87a6dcf12e57753/mmdet/models/utils/ckpt_convert.py#L96. +def reverse_correct_unfold_norm_order(x): + in_channel = x.shape[0] + x = x.reshape(in_channel // 4, 4) + x = x[:, [0, 2, 1, 3]].transpose(0, 1).reshape(in_channel) + return x + + +def convert_upernet_checkpoint(model_name, pytorch_dump_folder_path, push_to_hub): + model_name_to_url = { + "upernet-swin-tiny": "https://download.openmmlab.com/mmsegmentation/v0.5/swin/upernet_swin_tiny_patch4_window7_512x512_160k_ade20k_pretrain_224x224_1K/upernet_swin_tiny_patch4_window7_512x512_160k_ade20k_pretrain_224x224_1K_20210531_112542-e380ad3e.pth", + "upernet-swin-small": "https://download.openmmlab.com/mmsegmentation/v0.5/swin/upernet_swin_small_patch4_window7_512x512_160k_ade20k_pretrain_224x224_1K/upernet_swin_small_patch4_window7_512x512_160k_ade20k_pretrain_224x224_1K_20210526_192015-ee2fff1c.pth", + "upernet-swin-base": "https://download.openmmlab.com/mmsegmentation/v0.5/swin/upernet_swin_base_patch4_window12_512x512_160k_ade20k_pretrain_384x384_22K/upernet_swin_base_patch4_window12_512x512_160k_ade20k_pretrain_384x384_22K_20210531_125459-429057bf.pth", + "upernet-swin-large": "https://download.openmmlab.com/mmsegmentation/v0.5/swin/upernet_swin_large_patch4_window12_512x512_pretrain_384x384_22K_160k_ade20k/upernet_swin_large_patch4_window12_512x512_pretrain_384x384_22K_160k_ade20k_20220318_091743-9ba68901.pth", + } + checkpoint_url = model_name_to_url[model_name] + state_dict = torch.hub.load_state_dict_from_url(checkpoint_url, map_location="cpu", file_name=model_name)[ + "state_dict" + ] + + for name, param in state_dict.items(): + print(name, param.shape) + + config = get_upernet_config(model_name) + model = UperNetForSemanticSegmentation(config) + model.eval() + + # replace "bn" => "batch_norm" + for key in state_dict.copy().keys(): + val = state_dict.pop(key) + if "bn" in key: + key = key.replace("bn", "batch_norm") + state_dict[key] = val + + # rename keys + rename_keys = create_rename_keys(config) + for src, dest in rename_keys: + rename_key(state_dict, src, dest) + read_in_q_k_v(state_dict, config.backbone_config) + + # fix downsample parameters + for key, value in state_dict.items(): + if "downsample" in key: + if "reduction" in key: + state_dict[key] = reverse_correct_unfold_reduction_order(value) + if "norm" in key: + state_dict[key] = reverse_correct_unfold_norm_order(value) + + model.load_state_dict(state_dict) + + # verify on image + url = "https://huggingface.co/datasets/hf-internal-testing/fixtures_ade20k/resolve/main/ADE_val_00000001.jpg" + image = Image.open(requests.get(url, stream=True).raw).convert("RGB") + + processor = SegformerImageProcessor() + pixel_values = processor(image, return_tensors="pt").pixel_values + + with torch.no_grad(): + outputs = model(pixel_values) + logits = outputs.logits + + print(logits.shape) + print("First values of logits:", logits[0, 0, :3, :3]) + # assert values + if model_name == "upernet-swin-tiny": + expected_slice = torch.tensor( + [[-7.5958, -7.5958, -7.4302], [-7.5958, -7.5958, -7.4302], [-7.4797, -7.4797, -7.3068]] + ) + elif model_name == "upernet-swin-small": + expected_slice = torch.tensor( + [[-7.1921, -7.1921, -6.9532], [-7.1921, -7.1921, -6.9532], [-7.0908, -7.0908, -6.8534]] + ) + elif model_name == "upernet-swin-base": + expected_slice = torch.tensor( + [[-6.5851, -6.5851, -6.4330], [-6.5851, -6.5851, -6.4330], [-6.4763, -6.4763, -6.3254]] + ) + elif model_name == "upernet-swin-large": + expected_slice = torch.tensor( + [[-7.5297, -7.5297, -7.3802], [-7.5297, -7.5297, -7.3802], [-7.4044, -7.4044, -7.2586]] + ) + print("Logits:", outputs.logits[0, 0, :3, :3]) + assert torch.allclose(outputs.logits[0, 0, :3, :3], expected_slice, atol=1e-4) + print("Looks ok!") + + if pytorch_dump_folder_path is not None: + print(f"Saving model {model_name} to {pytorch_dump_folder_path}") + model.save_pretrained(pytorch_dump_folder_path) + print(f"Saving processor to {pytorch_dump_folder_path}") + processor.save_pretrained(pytorch_dump_folder_path) + + if push_to_hub: + print(f"Pushing model and processor for {model_name} to hub") + model.push_to_hub(f"openmmlab/{model_name}") + processor.push_to_hub(f"openmmlab/{model_name}") + + +if __name__ == "__main__": + parser = argparse.ArgumentParser() + # Required parameters + parser.add_argument( + "--model_name", + default="upernet-swin-tiny", + type=str, + choices=[f"upernet-swin-{size}" for size in ["tiny", "small", "base", "large"]], + help="Name of the Swin + UperNet model you'd like to convert.", + ) + parser.add_argument( + "--pytorch_dump_folder_path", default=None, type=str, help="Path to the output PyTorch model directory." + ) + parser.add_argument( + "--push_to_hub", action="store_true", help="Whether or not to push the converted model to the 🤗 hub." + ) + + args = parser.parse_args() + convert_upernet_checkpoint(args.model_name, args.pytorch_dump_folder_path, args.push_to_hub) diff --git a/venv/lib/python3.10/site-packages/transformers/models/upernet/modeling_upernet.py b/venv/lib/python3.10/site-packages/transformers/models/upernet/modeling_upernet.py new file mode 100644 index 0000000000000000000000000000000000000000..2d5b4443e35df3ddbf75744b6198915ae724bb6f --- /dev/null +++ b/venv/lib/python3.10/site-packages/transformers/models/upernet/modeling_upernet.py @@ -0,0 +1,440 @@ +# coding=utf-8 +# Copyright 2022 The HuggingFace Inc. team. All rights reserved. +# +# Licensed under the Apache License, Version 2.0 (the "License"); +# you may not use this file except in compliance with the License. +# You may obtain a copy of the License at +# +# http://www.apache.org/licenses/LICENSE-2.0 +# +# Unless required by applicable law or agreed to in writing, software +# distributed under the License is distributed on an "AS IS" BASIS, +# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. +# See the License for the specific language governing permissions and +# limitations under the License. +""" PyTorch UperNet model. Based on OpenMMLab's implementation, found in https://github.com/open-mmlab/mmsegmentation.""" + +from typing import List, Optional, Tuple, Union + +import torch +from torch import nn +from torch.nn import CrossEntropyLoss + +from ...modeling_outputs import SemanticSegmenterOutput +from ...modeling_utils import PreTrainedModel +from ...utils import add_start_docstrings, add_start_docstrings_to_model_forward, replace_return_docstrings +from ...utils.backbone_utils import load_backbone +from .configuration_upernet import UperNetConfig + + +# General docstring +_CONFIG_FOR_DOC = "UperNetConfig" + + +class UperNetConvModule(nn.Module): + """ + A convolutional block that bundles conv/norm/activation layers. This block simplifies the usage of convolution + layers, which are commonly used with a norm layer (e.g., BatchNorm) and activation layer (e.g., ReLU). + """ + + def __init__( + self, + in_channels: int, + out_channels: int, + kernel_size: Union[int, Tuple[int, int]], + padding: Union[int, Tuple[int, int], str] = 0, + bias: bool = False, + dilation: Union[int, Tuple[int, int]] = 1, + ) -> None: + super().__init__() + self.conv = nn.Conv2d( + in_channels=in_channels, + out_channels=out_channels, + kernel_size=kernel_size, + padding=padding, + bias=bias, + dilation=dilation, + ) + self.batch_norm = nn.BatchNorm2d(out_channels) + self.activation = nn.ReLU() + + def forward(self, input: torch.Tensor) -> torch.Tensor: + output = self.conv(input) + output = self.batch_norm(output) + output = self.activation(output) + + return output + + +class UperNetPyramidPoolingBlock(nn.Module): + def __init__(self, pool_scale: int, in_channels: int, channels: int) -> None: + super().__init__() + self.layers = [ + nn.AdaptiveAvgPool2d(pool_scale), + UperNetConvModule(in_channels, channels, kernel_size=1), + ] + for i, layer in enumerate(self.layers): + self.add_module(str(i), layer) + + def forward(self, input: torch.Tensor) -> torch.Tensor: + hidden_state = input + for layer in self.layers: + hidden_state = layer(hidden_state) + return hidden_state + + +class UperNetPyramidPoolingModule(nn.Module): + """ + Pyramid Pooling Module (PPM) used in PSPNet. + + Args: + pool_scales (`Tuple[int]`): + Pooling scales used in Pooling Pyramid Module. + in_channels (`int`): + Input channels. + channels (`int`): + Channels after modules, before conv_seg. + align_corners (`bool`): + align_corners argument of F.interpolate. + """ + + def __init__(self, pool_scales: Tuple[int, ...], in_channels: int, channels: int, align_corners: bool) -> None: + super().__init__() + self.pool_scales = pool_scales + self.align_corners = align_corners + self.in_channels = in_channels + self.channels = channels + self.blocks = [] + for i, pool_scale in enumerate(pool_scales): + block = UperNetPyramidPoolingBlock(pool_scale=pool_scale, in_channels=in_channels, channels=channels) + self.blocks.append(block) + self.add_module(str(i), block) + + def forward(self, x: torch.Tensor) -> List[torch.Tensor]: + ppm_outs = [] + for ppm in self.blocks: + ppm_out = ppm(x) + upsampled_ppm_out = nn.functional.interpolate( + ppm_out, size=x.size()[2:], mode="bilinear", align_corners=self.align_corners + ) + ppm_outs.append(upsampled_ppm_out) + return ppm_outs + + +class UperNetHead(nn.Module): + """ + Unified Perceptual Parsing for Scene Understanding. This head is the implementation of + [UPerNet](https://arxiv.org/abs/1807.10221). + """ + + def __init__(self, config, in_channels): + super().__init__() + + self.config = config + self.pool_scales = config.pool_scales # e.g. (1, 2, 3, 6) + self.in_channels = in_channels + self.channels = config.hidden_size + self.align_corners = False + self.classifier = nn.Conv2d(self.channels, config.num_labels, kernel_size=1) + + # PSP Module + self.psp_modules = UperNetPyramidPoolingModule( + self.pool_scales, + self.in_channels[-1], + self.channels, + align_corners=self.align_corners, + ) + self.bottleneck = UperNetConvModule( + self.in_channels[-1] + len(self.pool_scales) * self.channels, + self.channels, + kernel_size=3, + padding=1, + ) + # FPN Module + self.lateral_convs = nn.ModuleList() + self.fpn_convs = nn.ModuleList() + for in_channels in self.in_channels[:-1]: # skip the top layer + l_conv = UperNetConvModule(in_channels, self.channels, kernel_size=1) + fpn_conv = UperNetConvModule(self.channels, self.channels, kernel_size=3, padding=1) + self.lateral_convs.append(l_conv) + self.fpn_convs.append(fpn_conv) + + self.fpn_bottleneck = UperNetConvModule( + len(self.in_channels) * self.channels, + self.channels, + kernel_size=3, + padding=1, + ) + + def init_weights(self): + self.apply(self._init_weights) + + def _init_weights(self, module): + if isinstance(module, nn.Conv2d): + module.weight.data.normal_(mean=0.0, std=self.config.initializer_range) + if module.bias is not None: + module.bias.data.zero_() + + def psp_forward(self, inputs): + x = inputs[-1] + psp_outs = [x] + psp_outs.extend(self.psp_modules(x)) + psp_outs = torch.cat(psp_outs, dim=1) + output = self.bottleneck(psp_outs) + + return output + + def forward(self, encoder_hidden_states: torch.Tensor) -> torch.Tensor: + # build laterals + laterals = [lateral_conv(encoder_hidden_states[i]) for i, lateral_conv in enumerate(self.lateral_convs)] + + laterals.append(self.psp_forward(encoder_hidden_states)) + + # build top-down path + used_backbone_levels = len(laterals) + for i in range(used_backbone_levels - 1, 0, -1): + prev_shape = laterals[i - 1].shape[2:] + laterals[i - 1] = laterals[i - 1] + nn.functional.interpolate( + laterals[i], size=prev_shape, mode="bilinear", align_corners=self.align_corners + ) + + # build outputs + fpn_outs = [self.fpn_convs[i](laterals[i]) for i in range(used_backbone_levels - 1)] + # append psp feature + fpn_outs.append(laterals[-1]) + + for i in range(used_backbone_levels - 1, 0, -1): + fpn_outs[i] = nn.functional.interpolate( + fpn_outs[i], size=fpn_outs[0].shape[2:], mode="bilinear", align_corners=self.align_corners + ) + fpn_outs = torch.cat(fpn_outs, dim=1) + output = self.fpn_bottleneck(fpn_outs) + output = self.classifier(output) + + return output + + +class UperNetFCNHead(nn.Module): + """ + Fully Convolution Networks for Semantic Segmentation. This head is the implementation of + [FCNNet](https://arxiv.org/abs/1411.4038>). + + Args: + config: + Configuration. + in_channels (int): + Number of input channels. + kernel_size (int): + The kernel size for convs in the head. Default: 3. + dilation (int): + The dilation rate for convs in the head. Default: 1. + """ + + def __init__( + self, config, in_index: int = 2, kernel_size: int = 3, dilation: Union[int, Tuple[int, int]] = 1 + ) -> None: + super().__init__() + + self.config = config + self.in_channels = config.auxiliary_in_channels + self.channels = config.auxiliary_channels + self.num_convs = config.auxiliary_num_convs + self.concat_input = config.auxiliary_concat_input + self.in_index = in_index + + conv_padding = (kernel_size // 2) * dilation + convs = [] + convs.append( + UperNetConvModule( + self.in_channels, self.channels, kernel_size=kernel_size, padding=conv_padding, dilation=dilation + ) + ) + for i in range(self.num_convs - 1): + convs.append( + UperNetConvModule( + self.channels, self.channels, kernel_size=kernel_size, padding=conv_padding, dilation=dilation + ) + ) + if self.num_convs == 0: + self.convs = nn.Identity() + else: + self.convs = nn.Sequential(*convs) + if self.concat_input: + self.conv_cat = UperNetConvModule( + self.in_channels + self.channels, self.channels, kernel_size=kernel_size, padding=kernel_size // 2 + ) + + self.classifier = nn.Conv2d(self.channels, config.num_labels, kernel_size=1) + + def init_weights(self): + self.apply(self._init_weights) + + def _init_weights(self, module): + if isinstance(module, nn.Conv2d): + module.weight.data.normal_(mean=0.0, std=self.config.initializer_range) + if module.bias is not None: + module.bias.data.zero_() + + def forward(self, encoder_hidden_states: torch.Tensor) -> torch.Tensor: + # just take the relevant feature maps + hidden_states = encoder_hidden_states[self.in_index] + output = self.convs(hidden_states) + if self.concat_input: + output = self.conv_cat(torch.cat([hidden_states, output], dim=1)) + output = self.classifier(output) + return output + + +class UperNetPreTrainedModel(PreTrainedModel): + """ + An abstract class to handle weights initialization and a simple interface for downloading and loading pretrained + models. + """ + + config_class = UperNetConfig + main_input_name = "pixel_values" + + def _init_weights(self, module): + if isinstance(module, UperNetPreTrainedModel): + module.backbone.init_weights() + module.decode_head.init_weights() + if module.auxiliary_head is not None: + module.auxiliary_head.init_weights() + + def init_weights(self): + """Initialize the weights""" + self.backbone.init_weights() + self.decode_head.init_weights() + if self.auxiliary_head is not None: + self.auxiliary_head.init_weights() + + +UPERNET_START_DOCSTRING = r""" + Parameters: + This model is a PyTorch [torch.nn.Module](https://pytorch.org/docs/stable/nn.html#torch.nn.Module) sub-class. Use + it as a regular PyTorch Module and refer to the PyTorch documentation for all matter related to general usage and + behavior. + config ([`UperNetConfig`]): Model configuration class with all the parameters of the model. + Initializing with a config file does not load the weights associated with the model, only the + configuration. Check out the [`~PreTrainedModel.from_pretrained`] method to load the model weights. +""" + +UPERNET_INPUTS_DOCSTRING = r""" + Args: + pixel_values (`torch.FloatTensor` of shape `(batch_size, num_channels, height, width)`): + Pixel values. Padding will be ignored by default should you provide it. Pixel values can be obtained using + [`AutoImageProcessor`]. See [`SegformerImageProcessor.__call__`] for details. + output_attentions (`bool`, *optional*): + Whether or not to return the attentions tensors of all attention layers in case the backbone has them. See + `attentions` under returned tensors for more detail. + output_hidden_states (`bool`, *optional*): + Whether or not to return the hidden states of all layers of the backbone. See `hidden_states` under + returned tensors for more detail. + return_dict (`bool`, *optional*): + Whether or not to return a [`~utils.ModelOutput`] instead of a plain tuple. +""" + + +@add_start_docstrings( + """UperNet framework leveraging any vision backbone e.g. for ADE20k, CityScapes.""", + UPERNET_START_DOCSTRING, +) +class UperNetForSemanticSegmentation(UperNetPreTrainedModel): + def __init__(self, config): + super().__init__(config) + + self.backbone = load_backbone(config) + + # Semantic segmentation head(s) + self.decode_head = UperNetHead(config, in_channels=self.backbone.channels) + self.auxiliary_head = UperNetFCNHead(config) if config.use_auxiliary_head else None + + # Initialize weights and apply final processing + self.post_init() + + @add_start_docstrings_to_model_forward(UPERNET_INPUTS_DOCSTRING.format("batch_size, sequence_length")) + @replace_return_docstrings(output_type=SemanticSegmenterOutput, config_class=_CONFIG_FOR_DOC) + def forward( + self, + pixel_values: Optional[torch.Tensor] = None, + output_attentions: Optional[bool] = None, + output_hidden_states: Optional[bool] = None, + labels: Optional[torch.Tensor] = None, + return_dict: Optional[bool] = None, + ) -> Union[tuple, SemanticSegmenterOutput]: + r""" + labels (`torch.LongTensor` of shape `(batch_size, height, width)`, *optional*): + Ground truth semantic segmentation maps for computing the loss. Indices should be in `[0, ..., + config.num_labels - 1]`. If `config.num_labels > 1`, a classification loss is computed (Cross-Entropy). + + Returns: + + Examples: + ```python + >>> from transformers import AutoImageProcessor, UperNetForSemanticSegmentation + >>> from PIL import Image + >>> from huggingface_hub import hf_hub_download + + >>> image_processor = AutoImageProcessor.from_pretrained("openmmlab/upernet-convnext-tiny") + >>> model = UperNetForSemanticSegmentation.from_pretrained("openmmlab/upernet-convnext-tiny") + + >>> filepath = hf_hub_download( + ... repo_id="hf-internal-testing/fixtures_ade20k", filename="ADE_val_00000001.jpg", repo_type="dataset" + ... ) + >>> image = Image.open(filepath).convert("RGB") + + >>> inputs = image_processor(images=image, return_tensors="pt") + + >>> outputs = model(**inputs) + + >>> logits = outputs.logits # shape (batch_size, num_labels, height, width) + >>> list(logits.shape) + [1, 150, 512, 512] + ```""" + + return_dict = return_dict if return_dict is not None else self.config.use_return_dict + output_hidden_states = ( + output_hidden_states if output_hidden_states is not None else self.config.output_hidden_states + ) + output_attentions = output_attentions if output_attentions is not None else self.config.output_attentions + + outputs = self.backbone.forward_with_filtered_kwargs( + pixel_values, output_hidden_states=output_hidden_states, output_attentions=output_attentions + ) + features = outputs.feature_maps + + logits = self.decode_head(features) + logits = nn.functional.interpolate(logits, size=pixel_values.shape[2:], mode="bilinear", align_corners=False) + + auxiliary_logits = None + if self.auxiliary_head is not None: + auxiliary_logits = self.auxiliary_head(features) + auxiliary_logits = nn.functional.interpolate( + auxiliary_logits, size=pixel_values.shape[2:], mode="bilinear", align_corners=False + ) + + loss = None + if labels is not None: + if self.config.num_labels == 1: + raise ValueError("The number of labels should be greater than one") + else: + # compute weighted loss + loss_fct = CrossEntropyLoss(ignore_index=self.config.loss_ignore_index) + loss = loss_fct(logits, labels) + if auxiliary_logits is not None: + auxiliary_loss = loss_fct(auxiliary_logits, labels) + loss += self.config.auxiliary_loss_weight * auxiliary_loss + + if not return_dict: + if output_hidden_states: + output = (logits,) + outputs[1:] + else: + output = (logits,) + outputs[2:] + return ((loss,) + output) if loss is not None else output + + return SemanticSegmenterOutput( + loss=loss, + logits=logits, + hidden_states=outputs.hidden_states, + attentions=outputs.attentions, + ) diff --git a/venv/lib/python3.10/site-packages/transformers/models/yolos/__pycache__/__init__.cpython-310.pyc b/venv/lib/python3.10/site-packages/transformers/models/yolos/__pycache__/__init__.cpython-310.pyc new file mode 100644 index 0000000000000000000000000000000000000000..05853678f3dfd25c2329a7d967139d7f19cf66e9 Binary files /dev/null and b/venv/lib/python3.10/site-packages/transformers/models/yolos/__pycache__/__init__.cpython-310.pyc differ diff --git a/venv/lib/python3.10/site-packages/transformers/models/yolos/__pycache__/configuration_yolos.cpython-310.pyc b/venv/lib/python3.10/site-packages/transformers/models/yolos/__pycache__/configuration_yolos.cpython-310.pyc new file mode 100644 index 0000000000000000000000000000000000000000..f440901efe96467760c02032936a6ae712991201 Binary files /dev/null and b/venv/lib/python3.10/site-packages/transformers/models/yolos/__pycache__/configuration_yolos.cpython-310.pyc differ diff --git a/venv/lib/python3.10/site-packages/transformers/models/yolos/__pycache__/convert_yolos_to_pytorch.cpython-310.pyc b/venv/lib/python3.10/site-packages/transformers/models/yolos/__pycache__/convert_yolos_to_pytorch.cpython-310.pyc new file mode 100644 index 0000000000000000000000000000000000000000..a9be515143924eee7fb3997c1e86fcbc87e15fb8 Binary files /dev/null and b/venv/lib/python3.10/site-packages/transformers/models/yolos/__pycache__/convert_yolos_to_pytorch.cpython-310.pyc differ diff --git a/venv/lib/python3.10/site-packages/transformers/models/yolos/__pycache__/feature_extraction_yolos.cpython-310.pyc b/venv/lib/python3.10/site-packages/transformers/models/yolos/__pycache__/feature_extraction_yolos.cpython-310.pyc new file mode 100644 index 0000000000000000000000000000000000000000..f430005de5bcacae3c5d2e9f713fd95258b5318c Binary files /dev/null and b/venv/lib/python3.10/site-packages/transformers/models/yolos/__pycache__/feature_extraction_yolos.cpython-310.pyc differ diff --git a/venv/lib/python3.10/site-packages/transformers/models/yolos/__pycache__/image_processing_yolos.cpython-310.pyc b/venv/lib/python3.10/site-packages/transformers/models/yolos/__pycache__/image_processing_yolos.cpython-310.pyc new file mode 100644 index 0000000000000000000000000000000000000000..a9a4b313d9021a14d7b934472607d281063e86ac Binary files /dev/null and b/venv/lib/python3.10/site-packages/transformers/models/yolos/__pycache__/image_processing_yolos.cpython-310.pyc differ diff --git a/venv/lib/python3.10/site-packages/transformers/models/yolos/__pycache__/modeling_yolos.cpython-310.pyc b/venv/lib/python3.10/site-packages/transformers/models/yolos/__pycache__/modeling_yolos.cpython-310.pyc new file mode 100644 index 0000000000000000000000000000000000000000..3f7ca13523c297c77e20805053ce81fd04a21078 Binary files /dev/null and b/venv/lib/python3.10/site-packages/transformers/models/yolos/__pycache__/modeling_yolos.cpython-310.pyc differ diff --git a/venv/lib/python3.10/site-packages/transformers/models/yolos/configuration_yolos.py b/venv/lib/python3.10/site-packages/transformers/models/yolos/configuration_yolos.py new file mode 100644 index 0000000000000000000000000000000000000000..098210f1a732e28260fe92e2512d59745ad52195 --- /dev/null +++ b/venv/lib/python3.10/site-packages/transformers/models/yolos/configuration_yolos.py @@ -0,0 +1,178 @@ +# coding=utf-8 +# Copyright 2022 The HuggingFace Inc. team. All rights reserved. +# +# Licensed under the Apache License, Version 2.0 (the "License"); +# you may not use this file except in compliance with the License. +# You may obtain a copy of the License at +# +# http://www.apache.org/licenses/LICENSE-2.0 +# +# Unless required by applicable law or agreed to in writing, software +# distributed under the License is distributed on an "AS IS" BASIS, +# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. +# See the License for the specific language governing permissions and +# limitations under the License. +""" YOLOS model configuration""" + +from collections import OrderedDict +from typing import Mapping + +from packaging import version + +from ...configuration_utils import PretrainedConfig +from ...onnx import OnnxConfig +from ...utils import logging + + +logger = logging.get_logger(__name__) + + +from ..deprecated._archive_maps import YOLOS_PRETRAINED_CONFIG_ARCHIVE_MAP # noqa: F401, E402 + + +class YolosConfig(PretrainedConfig): + r""" + This is the configuration class to store the configuration of a [`YolosModel`]. It is used to instantiate a YOLOS + model according to the specified arguments, defining the model architecture. Instantiating a configuration with the + defaults will yield a similar configuration to that of the YOLOS + [hustvl/yolos-base](https://huggingface.co/hustvl/yolos-base) architecture. + + Configuration objects inherit from [`PretrainedConfig`] and can be used to control the model outputs. Read the + documentation from [`PretrainedConfig`] for more information. + + Args: + hidden_size (`int`, *optional*, defaults to 768): + Dimensionality of the encoder layers and the pooler layer. + num_hidden_layers (`int`, *optional*, defaults to 12): + Number of hidden layers in the Transformer encoder. + num_attention_heads (`int`, *optional*, defaults to 12): + Number of attention heads for each attention layer in the Transformer encoder. + intermediate_size (`int`, *optional*, defaults to 3072): + Dimensionality of the "intermediate" (i.e., feed-forward) layer in the Transformer encoder. + hidden_act (`str` or `function`, *optional*, defaults to `"gelu"`): + The non-linear activation function (function or string) in the encoder and pooler. If string, `"gelu"`, + `"relu"`, `"selu"` and `"gelu_new"` are supported. + hidden_dropout_prob (`float`, *optional*, defaults to 0.0): + The dropout probability for all fully connected layers in the embeddings, encoder, and pooler. + attention_probs_dropout_prob (`float`, *optional*, defaults to 0.0): + The dropout ratio for the attention probabilities. + initializer_range (`float`, *optional*, defaults to 0.02): + The standard deviation of the truncated_normal_initializer for initializing all weight matrices. + layer_norm_eps (`float`, *optional*, defaults to 1e-12): + The epsilon used by the layer normalization layers. + image_size (`List[int]`, *optional*, defaults to `[512, 864]`): + The size (resolution) of each image. + patch_size (`int`, *optional*, defaults to 16): + The size (resolution) of each patch. + num_channels (`int`, *optional*, defaults to 3): + The number of input channels. + qkv_bias (`bool`, *optional*, defaults to `True`): + Whether to add a bias to the queries, keys and values. + num_detection_tokens (`int`, *optional*, defaults to 100): + The number of detection tokens. + use_mid_position_embeddings (`bool`, *optional*, defaults to `True`): + Whether to use the mid-layer position encodings. + auxiliary_loss (`bool`, *optional*, defaults to `False`): + Whether auxiliary decoding losses (loss at each decoder layer) are to be used. + class_cost (`float`, *optional*, defaults to 1): + Relative weight of the classification error in the Hungarian matching cost. + bbox_cost (`float`, *optional*, defaults to 5): + Relative weight of the L1 error of the bounding box coordinates in the Hungarian matching cost. + giou_cost (`float`, *optional*, defaults to 2): + Relative weight of the generalized IoU loss of the bounding box in the Hungarian matching cost. + bbox_loss_coefficient (`float`, *optional*, defaults to 5): + Relative weight of the L1 bounding box loss in the object detection loss. + giou_loss_coefficient (`float`, *optional*, defaults to 2): + Relative weight of the generalized IoU loss in the object detection loss. + eos_coefficient (`float`, *optional*, defaults to 0.1): + Relative classification weight of the 'no-object' class in the object detection loss. + + Example: + + ```python + >>> from transformers import YolosConfig, YolosModel + + >>> # Initializing a YOLOS hustvl/yolos-base style configuration + >>> configuration = YolosConfig() + + >>> # Initializing a model (with random weights) from the hustvl/yolos-base style configuration + >>> model = YolosModel(configuration) + + >>> # Accessing the model configuration + >>> configuration = model.config + ```""" + + model_type = "yolos" + + def __init__( + self, + hidden_size=768, + num_hidden_layers=12, + num_attention_heads=12, + intermediate_size=3072, + hidden_act="gelu", + hidden_dropout_prob=0.0, + attention_probs_dropout_prob=0.0, + initializer_range=0.02, + layer_norm_eps=1e-12, + image_size=[512, 864], + patch_size=16, + num_channels=3, + qkv_bias=True, + num_detection_tokens=100, + use_mid_position_embeddings=True, + auxiliary_loss=False, + class_cost=1, + bbox_cost=5, + giou_cost=2, + bbox_loss_coefficient=5, + giou_loss_coefficient=2, + eos_coefficient=0.1, + **kwargs, + ): + super().__init__(**kwargs) + + self.hidden_size = hidden_size + self.num_hidden_layers = num_hidden_layers + self.num_attention_heads = num_attention_heads + self.intermediate_size = intermediate_size + self.hidden_act = hidden_act + self.hidden_dropout_prob = hidden_dropout_prob + self.attention_probs_dropout_prob = attention_probs_dropout_prob + self.initializer_range = initializer_range + self.layer_norm_eps = layer_norm_eps + self.image_size = image_size + self.patch_size = patch_size + self.num_channels = num_channels + self.qkv_bias = qkv_bias + self.num_detection_tokens = num_detection_tokens + self.use_mid_position_embeddings = use_mid_position_embeddings + self.auxiliary_loss = auxiliary_loss + # Hungarian matcher + self.class_cost = class_cost + self.bbox_cost = bbox_cost + self.giou_cost = giou_cost + # Loss coefficients + self.bbox_loss_coefficient = bbox_loss_coefficient + self.giou_loss_coefficient = giou_loss_coefficient + self.eos_coefficient = eos_coefficient + + +class YolosOnnxConfig(OnnxConfig): + torch_onnx_minimum_version = version.parse("1.11") + + @property + def inputs(self) -> Mapping[str, Mapping[int, str]]: + return OrderedDict( + [ + ("pixel_values", {0: "batch", 1: "num_channels", 2: "height", 3: "width"}), + ] + ) + + @property + def atol_for_validation(self) -> float: + return 1e-4 + + @property + def default_onnx_opset(self) -> int: + return 12 diff --git a/venv/lib/python3.10/site-packages/transformers/models/yolos/image_processing_yolos.py b/venv/lib/python3.10/site-packages/transformers/models/yolos/image_processing_yolos.py new file mode 100644 index 0000000000000000000000000000000000000000..c4e44854a0da4322044017f395e6ad400f92f6b7 --- /dev/null +++ b/venv/lib/python3.10/site-packages/transformers/models/yolos/image_processing_yolos.py @@ -0,0 +1,1447 @@ +# coding=utf-8 +# Copyright 2022 The HuggingFace Inc. team. All rights reserved. +# +# Licensed under the Apache License, Version 2.0 (the "License"); +# you may not use this file except in compliance with the License. +# You may obtain a copy of the License at +# +# http://www.apache.org/licenses/LICENSE-2.0 +# +# Unless required by applicable law or agreed to in writing, software +# distributed under the License is distributed on an "AS IS" BASIS, +# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. +# See the License for the specific language governing permissions and +# limitations under the License. +"""Image processor class for YOLOS.""" + +import pathlib +from typing import Any, Callable, Dict, Iterable, List, Optional, Set, Tuple, Union + +import numpy as np + +from ...feature_extraction_utils import BatchFeature +from ...image_processing_utils import BaseImageProcessor, get_size_dict +from ...image_transforms import ( + PaddingMode, + center_to_corners_format, + corners_to_center_format, + id_to_rgb, + pad, + rescale, + resize, + rgb_to_id, + to_channel_dimension_format, +) +from ...image_utils import ( + IMAGENET_DEFAULT_MEAN, + IMAGENET_DEFAULT_STD, + AnnotationFormat, + AnnotationType, + ChannelDimension, + ImageInput, + PILImageResampling, + get_image_size, + infer_channel_dimension_format, + is_scaled_image, + make_list_of_images, + to_numpy_array, + valid_images, + validate_annotations, + validate_kwargs, + validate_preprocess_arguments, +) +from ...utils import ( + TensorType, + is_flax_available, + is_jax_tensor, + is_scipy_available, + is_tf_available, + is_tf_tensor, + is_torch_available, + is_torch_tensor, + is_vision_available, + logging, +) + + +if is_torch_available(): + import torch + from torch import nn + + +if is_vision_available(): + import PIL + + +if is_scipy_available(): + import scipy.special + import scipy.stats + +logger = logging.get_logger(__name__) + +SUPPORTED_ANNOTATION_FORMATS = (AnnotationFormat.COCO_DETECTION, AnnotationFormat.COCO_PANOPTIC) + + +# Copied from transformers.models.detr.image_processing_detr.get_max_height_width +def get_max_height_width( + images: List[np.ndarray], input_data_format: Optional[Union[str, ChannelDimension]] = None +) -> List[int]: + """ + Get the maximum height and width across all images in a batch. + """ + if input_data_format is None: + input_data_format = infer_channel_dimension_format(images[0]) + + if input_data_format == ChannelDimension.FIRST: + _, max_height, max_width = max_across_indices([img.shape for img in images]) + elif input_data_format == ChannelDimension.LAST: + max_height, max_width, _ = max_across_indices([img.shape for img in images]) + else: + raise ValueError(f"Invalid channel dimension format: {input_data_format}") + return (max_height, max_width) + + +def get_size_with_aspect_ratio(image_size, size, max_size=None) -> Tuple[int, int]: + """ + Computes the output image size given the input image size and the desired output size. + + Args: + image_size (`Tuple[int, int]`): + The input image size. + size (`int`): + The desired output size. + max_size (`int`, *optional*): + The maximum allowed output size. + """ + height, width = image_size + if max_size is not None: + min_original_size = float(min((height, width))) + max_original_size = float(max((height, width))) + if max_original_size / min_original_size * size > max_size: + size = int(round(max_size * min_original_size / max_original_size)) + + if width < height and width != size: + height = int(size * height / width) + width = size + elif height < width and height != size: + width = int(size * width / height) + height = size + width_mod = np.mod(width, 16) + height_mod = np.mod(height, 16) + width = width - width_mod + height = height - height_mod + return (height, width) + + +# Copied from transformers.models.detr.image_processing_detr.get_resize_output_image_size +def get_resize_output_image_size( + input_image: np.ndarray, + size: Union[int, Tuple[int, int], List[int]], + max_size: Optional[int] = None, + input_data_format: Optional[Union[str, ChannelDimension]] = None, +) -> Tuple[int, int]: + """ + Computes the output image size given the input image size and the desired output size. If the desired output size + is a tuple or list, the output image size is returned as is. If the desired output size is an integer, the output + image size is computed by keeping the aspect ratio of the input image size. + + Args: + input_image (`np.ndarray`): + The image to resize. + size (`int` or `Tuple[int, int]` or `List[int]`): + The desired output size. + max_size (`int`, *optional*): + The maximum allowed output size. + input_data_format (`ChannelDimension` or `str`, *optional*): + The channel dimension format of the input image. If not provided, it will be inferred from the input image. + """ + image_size = get_image_size(input_image, input_data_format) + if isinstance(size, (list, tuple)): + return size + + return get_size_with_aspect_ratio(image_size, size, max_size) + + +# Copied from transformers.models.detr.image_processing_detr.get_numpy_to_framework_fn +def get_numpy_to_framework_fn(arr) -> Callable: + """ + Returns a function that converts a numpy array to the framework of the input array. + + Args: + arr (`np.ndarray`): The array to convert. + """ + if isinstance(arr, np.ndarray): + return np.array + if is_tf_available() and is_tf_tensor(arr): + import tensorflow as tf + + return tf.convert_to_tensor + if is_torch_available() and is_torch_tensor(arr): + import torch + + return torch.tensor + if is_flax_available() and is_jax_tensor(arr): + import jax.numpy as jnp + + return jnp.array + raise ValueError(f"Cannot convert arrays of type {type(arr)}") + + +# Copied from transformers.models.detr.image_processing_detr.safe_squeeze +def safe_squeeze(arr: np.ndarray, axis: Optional[int] = None) -> np.ndarray: + """ + Squeezes an array, but only if the axis specified has dim 1. + """ + if axis is None: + return arr.squeeze() + + try: + return arr.squeeze(axis=axis) + except ValueError: + return arr + + +# Copied from transformers.models.detr.image_processing_detr.normalize_annotation +def normalize_annotation(annotation: Dict, image_size: Tuple[int, int]) -> Dict: + image_height, image_width = image_size + norm_annotation = {} + for key, value in annotation.items(): + if key == "boxes": + boxes = value + boxes = corners_to_center_format(boxes) + boxes /= np.asarray([image_width, image_height, image_width, image_height], dtype=np.float32) + norm_annotation[key] = boxes + else: + norm_annotation[key] = value + return norm_annotation + + +# Copied from transformers.models.detr.image_processing_detr.max_across_indices +def max_across_indices(values: Iterable[Any]) -> List[Any]: + """ + Return the maximum value across all indices of an iterable of values. + """ + return [max(values_i) for values_i in zip(*values)] + + +# Copied from transformers.models.detr.image_processing_detr.make_pixel_mask +def make_pixel_mask( + image: np.ndarray, output_size: Tuple[int, int], input_data_format: Optional[Union[str, ChannelDimension]] = None +) -> np.ndarray: + """ + Make a pixel mask for the image, where 1 indicates a valid pixel and 0 indicates padding. + + Args: + image (`np.ndarray`): + Image to make the pixel mask for. + output_size (`Tuple[int, int]`): + Output size of the mask. + """ + input_height, input_width = get_image_size(image, channel_dim=input_data_format) + mask = np.zeros(output_size, dtype=np.int64) + mask[:input_height, :input_width] = 1 + return mask + + +# Copied from transformers.models.detr.image_processing_detr.convert_coco_poly_to_mask +def convert_coco_poly_to_mask(segmentations, height: int, width: int) -> np.ndarray: + """ + Convert a COCO polygon annotation to a mask. + + Args: + segmentations (`List[List[float]]`): + List of polygons, each polygon represented by a list of x-y coordinates. + height (`int`): + Height of the mask. + width (`int`): + Width of the mask. + """ + try: + from pycocotools import mask as coco_mask + except ImportError: + raise ImportError("Pycocotools is not installed in your environment.") + + masks = [] + for polygons in segmentations: + rles = coco_mask.frPyObjects(polygons, height, width) + mask = coco_mask.decode(rles) + if len(mask.shape) < 3: + mask = mask[..., None] + mask = np.asarray(mask, dtype=np.uint8) + mask = np.any(mask, axis=2) + masks.append(mask) + if masks: + masks = np.stack(masks, axis=0) + else: + masks = np.zeros((0, height, width), dtype=np.uint8) + + return masks + + +# Copied from transformers.models.detr.image_processing_detr.prepare_coco_detection_annotation +def prepare_coco_detection_annotation( + image, + target, + return_segmentation_masks: bool = False, + input_data_format: Optional[Union[ChannelDimension, str]] = None, +): + """ + Convert the target in COCO format into the format expected by DETR. + """ + image_height, image_width = get_image_size(image, channel_dim=input_data_format) + + image_id = target["image_id"] + image_id = np.asarray([image_id], dtype=np.int64) + + # Get all COCO annotations for the given image. + annotations = target["annotations"] + annotations = [obj for obj in annotations if "iscrowd" not in obj or obj["iscrowd"] == 0] + + classes = [obj["category_id"] for obj in annotations] + classes = np.asarray(classes, dtype=np.int64) + + # for conversion to coco api + area = np.asarray([obj["area"] for obj in annotations], dtype=np.float32) + iscrowd = np.asarray([obj["iscrowd"] if "iscrowd" in obj else 0 for obj in annotations], dtype=np.int64) + + boxes = [obj["bbox"] for obj in annotations] + # guard against no boxes via resizing + boxes = np.asarray(boxes, dtype=np.float32).reshape(-1, 4) + boxes[:, 2:] += boxes[:, :2] + boxes[:, 0::2] = boxes[:, 0::2].clip(min=0, max=image_width) + boxes[:, 1::2] = boxes[:, 1::2].clip(min=0, max=image_height) + + keep = (boxes[:, 3] > boxes[:, 1]) & (boxes[:, 2] > boxes[:, 0]) + + new_target = {} + new_target["image_id"] = image_id + new_target["class_labels"] = classes[keep] + new_target["boxes"] = boxes[keep] + new_target["area"] = area[keep] + new_target["iscrowd"] = iscrowd[keep] + new_target["orig_size"] = np.asarray([int(image_height), int(image_width)], dtype=np.int64) + + if annotations and "keypoints" in annotations[0]: + keypoints = [obj["keypoints"] for obj in annotations] + # Converting the filtered keypoints list to a numpy array + keypoints = np.asarray(keypoints, dtype=np.float32) + # Apply the keep mask here to filter the relevant annotations + keypoints = keypoints[keep] + num_keypoints = keypoints.shape[0] + keypoints = keypoints.reshape((-1, 3)) if num_keypoints else keypoints + new_target["keypoints"] = keypoints + + if return_segmentation_masks: + segmentation_masks = [obj["segmentation"] for obj in annotations] + masks = convert_coco_poly_to_mask(segmentation_masks, image_height, image_width) + new_target["masks"] = masks[keep] + + return new_target + + +# Copied from transformers.models.detr.image_processing_detr.masks_to_boxes +def masks_to_boxes(masks: np.ndarray) -> np.ndarray: + """ + Compute the bounding boxes around the provided panoptic segmentation masks. + + Args: + masks: masks in format `[number_masks, height, width]` where N is the number of masks + + Returns: + boxes: bounding boxes in format `[number_masks, 4]` in xyxy format + """ + if masks.size == 0: + return np.zeros((0, 4)) + + h, w = masks.shape[-2:] + y = np.arange(0, h, dtype=np.float32) + x = np.arange(0, w, dtype=np.float32) + # see https://github.com/pytorch/pytorch/issues/50276 + y, x = np.meshgrid(y, x, indexing="ij") + + x_mask = masks * np.expand_dims(x, axis=0) + x_max = x_mask.reshape(x_mask.shape[0], -1).max(-1) + x = np.ma.array(x_mask, mask=~(np.array(masks, dtype=bool))) + x_min = x.filled(fill_value=1e8) + x_min = x_min.reshape(x_min.shape[0], -1).min(-1) + + y_mask = masks * np.expand_dims(y, axis=0) + y_max = y_mask.reshape(x_mask.shape[0], -1).max(-1) + y = np.ma.array(y_mask, mask=~(np.array(masks, dtype=bool))) + y_min = y.filled(fill_value=1e8) + y_min = y_min.reshape(y_min.shape[0], -1).min(-1) + + return np.stack([x_min, y_min, x_max, y_max], 1) + + +# Copied from transformers.models.detr.image_processing_detr.prepare_coco_panoptic_annotation with DETR->YOLOS +def prepare_coco_panoptic_annotation( + image: np.ndarray, + target: Dict, + masks_path: Union[str, pathlib.Path], + return_masks: bool = True, + input_data_format: Union[ChannelDimension, str] = None, +) -> Dict: + """ + Prepare a coco panoptic annotation for YOLOS. + """ + image_height, image_width = get_image_size(image, channel_dim=input_data_format) + annotation_path = pathlib.Path(masks_path) / target["file_name"] + + new_target = {} + new_target["image_id"] = np.asarray([target["image_id"] if "image_id" in target else target["id"]], dtype=np.int64) + new_target["size"] = np.asarray([image_height, image_width], dtype=np.int64) + new_target["orig_size"] = np.asarray([image_height, image_width], dtype=np.int64) + + if "segments_info" in target: + masks = np.asarray(PIL.Image.open(annotation_path), dtype=np.uint32) + masks = rgb_to_id(masks) + + ids = np.array([segment_info["id"] for segment_info in target["segments_info"]]) + masks = masks == ids[:, None, None] + masks = masks.astype(np.uint8) + if return_masks: + new_target["masks"] = masks + new_target["boxes"] = masks_to_boxes(masks) + new_target["class_labels"] = np.array( + [segment_info["category_id"] for segment_info in target["segments_info"]], dtype=np.int64 + ) + new_target["iscrowd"] = np.asarray( + [segment_info["iscrowd"] for segment_info in target["segments_info"]], dtype=np.int64 + ) + new_target["area"] = np.asarray( + [segment_info["area"] for segment_info in target["segments_info"]], dtype=np.float32 + ) + + return new_target + + +# Copied from transformers.models.detr.image_processing_detr.get_segmentation_image +def get_segmentation_image( + masks: np.ndarray, input_size: Tuple, target_size: Tuple, stuff_equiv_classes, deduplicate=False +): + h, w = input_size + final_h, final_w = target_size + + m_id = scipy.special.softmax(masks.transpose(0, 1), -1) + + if m_id.shape[-1] == 0: + # We didn't detect any mask :( + m_id = np.zeros((h, w), dtype=np.int64) + else: + m_id = m_id.argmax(-1).reshape(h, w) + + if deduplicate: + # Merge the masks corresponding to the same stuff class + for equiv in stuff_equiv_classes.values(): + for eq_id in equiv: + m_id[m_id == eq_id] = equiv[0] + + seg_img = id_to_rgb(m_id) + seg_img = resize(seg_img, (final_w, final_h), resample=PILImageResampling.NEAREST) + return seg_img + + +# Copied from transformers.models.detr.image_processing_detr.get_mask_area +def get_mask_area(seg_img: np.ndarray, target_size: Tuple[int, int], n_classes: int) -> np.ndarray: + final_h, final_w = target_size + np_seg_img = seg_img.astype(np.uint8) + np_seg_img = np_seg_img.reshape(final_h, final_w, 3) + m_id = rgb_to_id(np_seg_img) + area = [(m_id == i).sum() for i in range(n_classes)] + return area + + +# Copied from transformers.models.detr.image_processing_detr.score_labels_from_class_probabilities +def score_labels_from_class_probabilities(logits: np.ndarray) -> Tuple[np.ndarray, np.ndarray]: + probs = scipy.special.softmax(logits, axis=-1) + labels = probs.argmax(-1, keepdims=True) + scores = np.take_along_axis(probs, labels, axis=-1) + scores, labels = scores.squeeze(-1), labels.squeeze(-1) + return scores, labels + + +# Copied from transformers.models.detr.image_processing_detr.resize_annotation +def resize_annotation( + annotation: Dict[str, Any], + orig_size: Tuple[int, int], + target_size: Tuple[int, int], + threshold: float = 0.5, + resample: PILImageResampling = PILImageResampling.NEAREST, +): + """ + Resizes an annotation to a target size. + + Args: + annotation (`Dict[str, Any]`): + The annotation dictionary. + orig_size (`Tuple[int, int]`): + The original size of the input image. + target_size (`Tuple[int, int]`): + The target size of the image, as returned by the preprocessing `resize` step. + threshold (`float`, *optional*, defaults to 0.5): + The threshold used to binarize the segmentation masks. + resample (`PILImageResampling`, defaults to `PILImageResampling.NEAREST`): + The resampling filter to use when resizing the masks. + """ + ratios = tuple(float(s) / float(s_orig) for s, s_orig in zip(target_size, orig_size)) + ratio_height, ratio_width = ratios + + new_annotation = {} + new_annotation["size"] = target_size + + for key, value in annotation.items(): + if key == "boxes": + boxes = value + scaled_boxes = boxes * np.asarray([ratio_width, ratio_height, ratio_width, ratio_height], dtype=np.float32) + new_annotation["boxes"] = scaled_boxes + elif key == "area": + area = value + scaled_area = area * (ratio_width * ratio_height) + new_annotation["area"] = scaled_area + elif key == "masks": + masks = value[:, None] + masks = np.array([resize(mask, target_size, resample=resample) for mask in masks]) + masks = masks.astype(np.float32) + masks = masks[:, 0] > threshold + new_annotation["masks"] = masks + elif key == "size": + new_annotation["size"] = target_size + else: + new_annotation[key] = value + + return new_annotation + + +# Copied from transformers.models.detr.image_processing_detr.binary_mask_to_rle +def binary_mask_to_rle(mask): + """ + Converts given binary mask of shape `(height, width)` to the run-length encoding (RLE) format. + + Args: + mask (`torch.Tensor` or `numpy.array`): + A binary mask tensor of shape `(height, width)` where 0 denotes background and 1 denotes the target + segment_id or class_id. + Returns: + `List`: Run-length encoded list of the binary mask. Refer to COCO API for more information about the RLE + format. + """ + if is_torch_tensor(mask): + mask = mask.numpy() + + pixels = mask.flatten() + pixels = np.concatenate([[0], pixels, [0]]) + runs = np.where(pixels[1:] != pixels[:-1])[0] + 1 + runs[1::2] -= runs[::2] + return list(runs) + + +# Copied from transformers.models.detr.image_processing_detr.convert_segmentation_to_rle +def convert_segmentation_to_rle(segmentation): + """ + Converts given segmentation map of shape `(height, width)` to the run-length encoding (RLE) format. + + Args: + segmentation (`torch.Tensor` or `numpy.array`): + A segmentation map of shape `(height, width)` where each value denotes a segment or class id. + Returns: + `List[List]`: A list of lists, where each list is the run-length encoding of a segment / class id. + """ + segment_ids = torch.unique(segmentation) + + run_length_encodings = [] + for idx in segment_ids: + mask = torch.where(segmentation == idx, 1, 0) + rle = binary_mask_to_rle(mask) + run_length_encodings.append(rle) + + return run_length_encodings + + +# Copied from transformers.models.detr.image_processing_detr.remove_low_and_no_objects +def remove_low_and_no_objects(masks, scores, labels, object_mask_threshold, num_labels): + """ + Binarize the given masks using `object_mask_threshold`, it returns the associated values of `masks`, `scores` and + `labels`. + + Args: + masks (`torch.Tensor`): + A tensor of shape `(num_queries, height, width)`. + scores (`torch.Tensor`): + A tensor of shape `(num_queries)`. + labels (`torch.Tensor`): + A tensor of shape `(num_queries)`. + object_mask_threshold (`float`): + A number between 0 and 1 used to binarize the masks. + Raises: + `ValueError`: Raised when the first dimension doesn't match in all input tensors. + Returns: + `Tuple[`torch.Tensor`, `torch.Tensor`, `torch.Tensor`]`: The `masks`, `scores` and `labels` without the region + < `object_mask_threshold`. + """ + if not (masks.shape[0] == scores.shape[0] == labels.shape[0]): + raise ValueError("mask, scores and labels must have the same shape!") + + to_keep = labels.ne(num_labels) & (scores > object_mask_threshold) + + return masks[to_keep], scores[to_keep], labels[to_keep] + + +# Copied from transformers.models.detr.image_processing_detr.check_segment_validity +def check_segment_validity(mask_labels, mask_probs, k, mask_threshold=0.5, overlap_mask_area_threshold=0.8): + # Get the mask associated with the k class + mask_k = mask_labels == k + mask_k_area = mask_k.sum() + + # Compute the area of all the stuff in query k + original_area = (mask_probs[k] >= mask_threshold).sum() + mask_exists = mask_k_area > 0 and original_area > 0 + + # Eliminate disconnected tiny segments + if mask_exists: + area_ratio = mask_k_area / original_area + if not area_ratio.item() > overlap_mask_area_threshold: + mask_exists = False + + return mask_exists, mask_k + + +# Copied from transformers.models.detr.image_processing_detr.compute_segments +def compute_segments( + mask_probs, + pred_scores, + pred_labels, + mask_threshold: float = 0.5, + overlap_mask_area_threshold: float = 0.8, + label_ids_to_fuse: Optional[Set[int]] = None, + target_size: Tuple[int, int] = None, +): + height = mask_probs.shape[1] if target_size is None else target_size[0] + width = mask_probs.shape[2] if target_size is None else target_size[1] + + segmentation = torch.zeros((height, width), dtype=torch.int32, device=mask_probs.device) + segments: List[Dict] = [] + + if target_size is not None: + mask_probs = nn.functional.interpolate( + mask_probs.unsqueeze(0), size=target_size, mode="bilinear", align_corners=False + )[0] + + current_segment_id = 0 + + # Weigh each mask by its prediction score + mask_probs *= pred_scores.view(-1, 1, 1) + mask_labels = mask_probs.argmax(0) # [height, width] + + # Keep track of instances of each class + stuff_memory_list: Dict[str, int] = {} + for k in range(pred_labels.shape[0]): + pred_class = pred_labels[k].item() + should_fuse = pred_class in label_ids_to_fuse + + # Check if mask exists and large enough to be a segment + mask_exists, mask_k = check_segment_validity( + mask_labels, mask_probs, k, mask_threshold, overlap_mask_area_threshold + ) + + if mask_exists: + if pred_class in stuff_memory_list: + current_segment_id = stuff_memory_list[pred_class] + else: + current_segment_id += 1 + + # Add current object segment to final segmentation map + segmentation[mask_k] = current_segment_id + segment_score = round(pred_scores[k].item(), 6) + segments.append( + { + "id": current_segment_id, + "label_id": pred_class, + "was_fused": should_fuse, + "score": segment_score, + } + ) + if should_fuse: + stuff_memory_list[pred_class] = current_segment_id + + return segmentation, segments + + +class YolosImageProcessor(BaseImageProcessor): + r""" + Constructs a Detr image processor. + + Args: + format (`str`, *optional*, defaults to `"coco_detection"`): + Data format of the annotations. One of "coco_detection" or "coco_panoptic". + do_resize (`bool`, *optional*, defaults to `True`): + Controls whether to resize the image's (height, width) dimensions to the specified `size`. Can be + overridden by the `do_resize` parameter in the `preprocess` method. + size (`Dict[str, int]` *optional*, defaults to `{"shortest_edge": 800, "longest_edge": 1333}`): + Size of the image's (height, width) dimensions after resizing. Can be overridden by the `size` parameter in + the `preprocess` method. + resample (`PILImageResampling`, *optional*, defaults to `PILImageResampling.BILINEAR`): + Resampling filter to use if resizing the image. + do_rescale (`bool`, *optional*, defaults to `True`): + Controls whether to rescale the image by the specified scale `rescale_factor`. Can be overridden by the + `do_rescale` parameter in the `preprocess` method. + rescale_factor (`int` or `float`, *optional*, defaults to `1/255`): + Scale factor to use if rescaling the image. Can be overridden by the `rescale_factor` parameter in the + `preprocess` method. + do_normalize: + Controls whether to normalize the image. Can be overridden by the `do_normalize` parameter in the + `preprocess` method. + image_mean (`float` or `List[float]`, *optional*, defaults to `IMAGENET_DEFAULT_MEAN`): + Mean values to use when normalizing the image. Can be a single value or a list of values, one for each + channel. Can be overridden by the `image_mean` parameter in the `preprocess` method. + image_std (`float` or `List[float]`, *optional*, defaults to `IMAGENET_DEFAULT_STD`): + Standard deviation values to use when normalizing the image. Can be a single value or a list of values, one + for each channel. Can be overridden by the `image_std` parameter in the `preprocess` method. + do_pad (`bool`, *optional*, defaults to `True`): + Controls whether to pad the image. Can be overridden by the `do_pad` parameter in the `preprocess` + method. If `True` will pad the images in the batch to the largest height and width in the batch. + Padding will be applied to the bottom and right of the image with zeros. + """ + + model_input_names = ["pixel_values", "pixel_mask"] + + def __init__( + self, + format: Union[str, AnnotationFormat] = AnnotationFormat.COCO_DETECTION, + do_resize: bool = True, + size: Dict[str, int] = None, + resample: PILImageResampling = PILImageResampling.BILINEAR, + do_rescale: bool = True, + rescale_factor: Union[int, float] = 1 / 255, + do_normalize: bool = True, + image_mean: Union[float, List[float]] = None, + image_std: Union[float, List[float]] = None, + do_convert_annotations: Optional[bool] = None, + do_pad: bool = True, + **kwargs, + ) -> None: + if "pad_and_return_pixel_mask" in kwargs: + do_pad = kwargs.pop("pad_and_return_pixel_mask") + + if "max_size" in kwargs: + logger.warning_once( + "The `max_size` parameter is deprecated and will be removed in v4.26. " + "Please specify in `size['longest_edge'] instead`.", + ) + max_size = kwargs.pop("max_size") + else: + max_size = None if size is None else 1333 + + size = size if size is not None else {"shortest_edge": 800, "longest_edge": 1333} + size = get_size_dict(size, max_size=max_size, default_to_square=False) + + # Backwards compatibility + if do_convert_annotations is None: + do_convert_annotations = do_normalize + + super().__init__(**kwargs) + self.format = format + self.do_resize = do_resize + self.size = size + self.resample = resample + self.do_rescale = do_rescale + self.rescale_factor = rescale_factor + self.do_normalize = do_normalize + self.do_convert_annotations = do_convert_annotations + self.image_mean = image_mean if image_mean is not None else IMAGENET_DEFAULT_MEAN + self.image_std = image_std if image_std is not None else IMAGENET_DEFAULT_STD + self.do_pad = do_pad + self._valid_processor_keys = [ + "images", + "annotations", + "return_segmentation_masks", + "masks_path", + "do_resize", + "size", + "resample", + "do_rescale", + "rescale_factor", + "do_normalize", + "image_mean", + "image_std", + "do_convert_annotations", + "do_pad", + "format", + "return_tensors", + "data_format", + "input_data_format", + ] + + @classmethod + # Copied from transformers.models.detr.image_processing_detr.DetrImageProcessor.from_dict with Detr->Yolos + def from_dict(cls, image_processor_dict: Dict[str, Any], **kwargs): + """ + Overrides the `from_dict` method from the base class to make sure parameters are updated if image processor is + created using from_dict and kwargs e.g. `YolosImageProcessor.from_pretrained(checkpoint, size=600, + max_size=800)` + """ + image_processor_dict = image_processor_dict.copy() + if "max_size" in kwargs: + image_processor_dict["max_size"] = kwargs.pop("max_size") + if "pad_and_return_pixel_mask" in kwargs: + image_processor_dict["pad_and_return_pixel_mask"] = kwargs.pop("pad_and_return_pixel_mask") + return super().from_dict(image_processor_dict, **kwargs) + + # Copied from transformers.models.detr.image_processing_detr.DetrImageProcessor.prepare_annotation + def prepare_annotation( + self, + image: np.ndarray, + target: Dict, + format: Optional[AnnotationFormat] = None, + return_segmentation_masks: bool = None, + masks_path: Optional[Union[str, pathlib.Path]] = None, + input_data_format: Optional[Union[str, ChannelDimension]] = None, + ) -> Dict: + """ + Prepare an annotation for feeding into DETR model. + """ + format = format if format is not None else self.format + + if format == AnnotationFormat.COCO_DETECTION: + return_segmentation_masks = False if return_segmentation_masks is None else return_segmentation_masks + target = prepare_coco_detection_annotation( + image, target, return_segmentation_masks, input_data_format=input_data_format + ) + elif format == AnnotationFormat.COCO_PANOPTIC: + return_segmentation_masks = True if return_segmentation_masks is None else return_segmentation_masks + target = prepare_coco_panoptic_annotation( + image, + target, + masks_path=masks_path, + return_masks=return_segmentation_masks, + input_data_format=input_data_format, + ) + else: + raise ValueError(f"Format {format} is not supported.") + return target + + # Copied from transformers.models.detr.image_processing_detr.DetrImageProcessor.prepare + def prepare(self, image, target, return_segmentation_masks=None, masks_path=None): + logger.warning_once( + "The `prepare` method is deprecated and will be removed in a v4.33. " + "Please use `prepare_annotation` instead. Note: the `prepare_annotation` method " + "does not return the image anymore.", + ) + target = self.prepare_annotation(image, target, return_segmentation_masks, masks_path, self.format) + return image, target + + # Copied from transformers.models.detr.image_processing_detr.DetrImageProcessor.convert_coco_poly_to_mask + def convert_coco_poly_to_mask(self, *args, **kwargs): + logger.warning_once("The `convert_coco_poly_to_mask` method is deprecated and will be removed in v4.33. ") + return convert_coco_poly_to_mask(*args, **kwargs) + + # Copied from transformers.models.detr.image_processing_detr.DetrImageProcessor.prepare_coco_detection with DETR->Yolos + def prepare_coco_detection(self, *args, **kwargs): + logger.warning_once("The `prepare_coco_detection` method is deprecated and will be removed in v4.33. ") + return prepare_coco_detection_annotation(*args, **kwargs) + + # Copied from transformers.models.detr.image_processing_detr.DetrImageProcessor.prepare_coco_panoptic + def prepare_coco_panoptic(self, *args, **kwargs): + logger.warning_once("The `prepare_coco_panoptic` method is deprecated and will be removed in v4.33. ") + return prepare_coco_panoptic_annotation(*args, **kwargs) + + # Copied from transformers.models.detr.image_processing_detr.DetrImageProcessor.resize + def resize( + self, + image: np.ndarray, + size: Dict[str, int], + resample: PILImageResampling = PILImageResampling.BILINEAR, + data_format: Optional[ChannelDimension] = None, + input_data_format: Optional[Union[str, ChannelDimension]] = None, + **kwargs, + ) -> np.ndarray: + """ + Resize the image to the given size. Size can be `min_size` (scalar) or `(height, width)` tuple. If size is an + int, smaller edge of the image will be matched to this number. + + Args: + image (`np.ndarray`): + Image to resize. + size (`Dict[str, int]`): + Dictionary containing the size to resize to. Can contain the keys `shortest_edge` and `longest_edge` or + `height` and `width`. + resample (`PILImageResampling`, *optional*, defaults to `PILImageResampling.BILINEAR`): + Resampling filter to use if resizing the image. + data_format (`str` or `ChannelDimension`, *optional*): + The channel dimension format for the output image. If unset, the channel dimension format of the input + image is used. + input_data_format (`ChannelDimension` or `str`, *optional*): + The channel dimension format of the input image. If not provided, it will be inferred. + """ + if "max_size" in kwargs: + logger.warning_once( + "The `max_size` parameter is deprecated and will be removed in v4.26. " + "Please specify in `size['longest_edge'] instead`.", + ) + max_size = kwargs.pop("max_size") + else: + max_size = None + size = get_size_dict(size, max_size=max_size, default_to_square=False) + if "shortest_edge" in size and "longest_edge" in size: + size = get_resize_output_image_size( + image, size["shortest_edge"], size["longest_edge"], input_data_format=input_data_format + ) + elif "height" in size and "width" in size: + size = (size["height"], size["width"]) + else: + raise ValueError( + "Size must contain 'height' and 'width' keys or 'shortest_edge' and 'longest_edge' keys. Got" + f" {size.keys()}." + ) + image = resize( + image, size=size, resample=resample, data_format=data_format, input_data_format=input_data_format, **kwargs + ) + return image + + # Copied from transformers.models.detr.image_processing_detr.DetrImageProcessor.resize_annotation + def resize_annotation( + self, + annotation, + orig_size, + size, + resample: PILImageResampling = PILImageResampling.NEAREST, + ) -> Dict: + """ + Resize the annotation to match the resized image. If size is an int, smaller edge of the mask will be matched + to this number. + """ + return resize_annotation(annotation, orig_size=orig_size, target_size=size, resample=resample) + + # Copied from transformers.models.detr.image_processing_detr.DetrImageProcessor.rescale + def rescale( + self, + image: np.ndarray, + rescale_factor: float, + data_format: Optional[Union[str, ChannelDimension]] = None, + input_data_format: Optional[Union[str, ChannelDimension]] = None, + ) -> np.ndarray: + """ + Rescale the image by the given factor. image = image * rescale_factor. + + Args: + image (`np.ndarray`): + Image to rescale. + rescale_factor (`float`): + The value to use for rescaling. + data_format (`str` or `ChannelDimension`, *optional*): + The channel dimension format for the output image. If unset, the channel dimension format of the input + image is used. Can be one of: + - `"channels_first"` or `ChannelDimension.FIRST`: image in (num_channels, height, width) format. + - `"channels_last"` or `ChannelDimension.LAST`: image in (height, width, num_channels) format. + input_data_format (`str` or `ChannelDimension`, *optional*): + The channel dimension format for the input image. If unset, is inferred from the input image. Can be + one of: + - `"channels_first"` or `ChannelDimension.FIRST`: image in (num_channels, height, width) format. + - `"channels_last"` or `ChannelDimension.LAST`: image in (height, width, num_channels) format. + """ + return rescale(image, rescale_factor, data_format=data_format, input_data_format=input_data_format) + + # Copied from transformers.models.detr.image_processing_detr.DetrImageProcessor.normalize_annotation + def normalize_annotation(self, annotation: Dict, image_size: Tuple[int, int]) -> Dict: + """ + Normalize the boxes in the annotation from `[top_left_x, top_left_y, bottom_right_x, bottom_right_y]` to + `[center_x, center_y, width, height]` format and from absolute to relative pixel values. + """ + return normalize_annotation(annotation, image_size=image_size) + + # Copied from transformers.models.detr.image_processing_detr.DetrImageProcessor._update_annotation_for_padded_image + def _update_annotation_for_padded_image( + self, + annotation: Dict, + input_image_size: Tuple[int, int], + output_image_size: Tuple[int, int], + padding, + update_bboxes, + ) -> Dict: + """ + Update the annotation for a padded image. + """ + new_annotation = {} + new_annotation["size"] = output_image_size + + for key, value in annotation.items(): + if key == "masks": + masks = value + masks = pad( + masks, + padding, + mode=PaddingMode.CONSTANT, + constant_values=0, + input_data_format=ChannelDimension.FIRST, + ) + masks = safe_squeeze(masks, 1) + new_annotation["masks"] = masks + elif key == "boxes" and update_bboxes: + boxes = value + boxes *= np.asarray( + [ + input_image_size[1] / output_image_size[1], + input_image_size[0] / output_image_size[0], + input_image_size[1] / output_image_size[1], + input_image_size[0] / output_image_size[0], + ] + ) + new_annotation["boxes"] = boxes + elif key == "size": + new_annotation["size"] = output_image_size + else: + new_annotation[key] = value + return new_annotation + + # Copied from transformers.models.detr.image_processing_detr.DetrImageProcessor._pad_image + def _pad_image( + self, + image: np.ndarray, + output_size: Tuple[int, int], + annotation: Optional[Dict[str, Any]] = None, + constant_values: Union[float, Iterable[float]] = 0, + data_format: Optional[ChannelDimension] = None, + input_data_format: Optional[Union[str, ChannelDimension]] = None, + update_bboxes: bool = True, + ) -> np.ndarray: + """ + Pad an image with zeros to the given size. + """ + input_height, input_width = get_image_size(image, channel_dim=input_data_format) + output_height, output_width = output_size + + pad_bottom = output_height - input_height + pad_right = output_width - input_width + padding = ((0, pad_bottom), (0, pad_right)) + padded_image = pad( + image, + padding, + mode=PaddingMode.CONSTANT, + constant_values=constant_values, + data_format=data_format, + input_data_format=input_data_format, + ) + if annotation is not None: + annotation = self._update_annotation_for_padded_image( + annotation, (input_height, input_width), (output_height, output_width), padding, update_bboxes + ) + return padded_image, annotation + + def pad( + self, + images: List[np.ndarray], + annotations: Optional[List[Dict[str, Any]]] = None, + constant_values: Union[float, Iterable[float]] = 0, + return_pixel_mask: bool = False, + return_tensors: Optional[Union[str, TensorType]] = None, + data_format: Optional[ChannelDimension] = None, + input_data_format: Optional[Union[str, ChannelDimension]] = None, + update_bboxes: bool = True, + ) -> BatchFeature: + """ + Pads a batch of images to the bottom and right of the image with zeros to the size of largest height and width + in the batch and optionally returns their corresponding pixel mask. + + Args: + image (`np.ndarray`): + Image to pad. + annotations (`List[Dict[str, any]]`, *optional*): + Annotations to pad along with the images. If provided, the bounding boxes will be updated to match the + padded images. + constant_values (`float` or `Iterable[float]`, *optional*): + The value to use for the padding if `mode` is `"constant"`. + return_pixel_mask (`bool`, *optional*, defaults to `True`): + Whether to return a pixel mask. + return_tensors (`str` or `TensorType`, *optional*): + The type of tensors to return. Can be one of: + - Unset: Return a list of `np.ndarray`. + - `TensorType.TENSORFLOW` or `'tf'`: Return a batch of type `tf.Tensor`. + - `TensorType.PYTORCH` or `'pt'`: Return a batch of type `torch.Tensor`. + - `TensorType.NUMPY` or `'np'`: Return a batch of type `np.ndarray`. + - `TensorType.JAX` or `'jax'`: Return a batch of type `jax.numpy.ndarray`. + data_format (`str` or `ChannelDimension`, *optional*): + The channel dimension format of the image. If not provided, it will be the same as the input image. + input_data_format (`ChannelDimension` or `str`, *optional*): + The channel dimension format of the input image. If not provided, it will be inferred. + update_bboxes (`bool`, *optional*, defaults to `True`): + Whether to update the bounding boxes in the annotations to match the padded images. If the + bounding boxes have not been converted to relative coordinates and `(centre_x, centre_y, width, height)` + format, the bounding boxes will not be updated. + """ + pad_size = get_max_height_width(images, input_data_format=input_data_format) + + annotation_list = annotations if annotations is not None else [None] * len(images) + padded_images = [] + padded_annotations = [] + for image, annotation in zip(images, annotation_list): + padded_image, padded_annotation = self._pad_image( + image, + pad_size, + annotation, + constant_values=constant_values, + data_format=data_format, + input_data_format=input_data_format, + update_bboxes=update_bboxes, + ) + padded_images.append(padded_image) + padded_annotations.append(padded_annotation) + + data = {"pixel_values": padded_images} + + if return_pixel_mask: + masks = [ + make_pixel_mask(image=image, output_size=pad_size, input_data_format=input_data_format) + for image in images + ] + data["pixel_mask"] = masks + + encoded_inputs = BatchFeature(data=data, tensor_type=return_tensors) + + if annotations is not None: + encoded_inputs["labels"] = [ + BatchFeature(annotation, tensor_type=return_tensors) for annotation in padded_annotations + ] + + return encoded_inputs + + def preprocess( + self, + images: ImageInput, + annotations: Optional[Union[AnnotationType, List[AnnotationType]]] = None, + return_segmentation_masks: bool = None, + masks_path: Optional[Union[str, pathlib.Path]] = None, + do_resize: Optional[bool] = None, + size: Optional[Dict[str, int]] = None, + resample=None, # PILImageResampling + do_rescale: Optional[bool] = None, + rescale_factor: Optional[Union[int, float]] = None, + do_normalize: Optional[bool] = None, + image_mean: Optional[Union[float, List[float]]] = None, + image_std: Optional[Union[float, List[float]]] = None, + do_convert_annotations: Optional[bool] = None, + do_pad: Optional[bool] = None, + format: Optional[Union[str, AnnotationFormat]] = None, + return_tensors: Optional[Union[TensorType, str]] = None, + data_format: Union[str, ChannelDimension] = ChannelDimension.FIRST, + input_data_format: Optional[Union[str, ChannelDimension]] = None, + **kwargs, + ) -> BatchFeature: + """ + Preprocess an image or a batch of images so that it can be used by the model. + + Args: + images (`ImageInput`): + Image or batch of images to preprocess. Expects a single or batch of images with pixel values ranging + from 0 to 255. If passing in images with pixel values between 0 and 1, set `do_rescale=False`. + annotations (`AnnotationType` or `List[AnnotationType]`, *optional*): + List of annotations associated with the image or batch of images. If annotation is for object + detection, the annotations should be a dictionary with the following keys: + - "image_id" (`int`): The image id. + - "annotations" (`List[Dict]`): List of annotations for an image. Each annotation should be a + dictionary. An image can have no annotations, in which case the list should be empty. + If annotation is for segmentation, the annotations should be a dictionary with the following keys: + - "image_id" (`int`): The image id. + - "segments_info" (`List[Dict]`): List of segments for an image. Each segment should be a dictionary. + An image can have no segments, in which case the list should be empty. + - "file_name" (`str`): The file name of the image. + return_segmentation_masks (`bool`, *optional*, defaults to self.return_segmentation_masks): + Whether to return segmentation masks. + masks_path (`str` or `pathlib.Path`, *optional*): + Path to the directory containing the segmentation masks. + do_resize (`bool`, *optional*, defaults to self.do_resize): + Whether to resize the image. + size (`Dict[str, int]`, *optional*, defaults to self.size): + Size of the image after resizing. + resample (`PILImageResampling`, *optional*, defaults to self.resample): + Resampling filter to use when resizing the image. + do_rescale (`bool`, *optional*, defaults to self.do_rescale): + Whether to rescale the image. + rescale_factor (`float`, *optional*, defaults to self.rescale_factor): + Rescale factor to use when rescaling the image. + do_normalize (`bool`, *optional*, defaults to self.do_normalize): + Whether to normalize the image. + image_mean (`float` or `List[float]`, *optional*, defaults to self.image_mean): + Mean to use when normalizing the image. + image_std (`float` or `List[float]`, *optional*, defaults to self.image_std): + Standard deviation to use when normalizing the image. + do_convert_annotations (`bool`, *optional*, defaults to self.do_convert_annotations): + Whether to convert the annotations to the format expected by the model. Converts the bounding + boxes from the format `(top_left_x, top_left_y, width, height)` to `(center_x, center_y, width, height)` + and in relative coordinates. + do_pad (`bool`, *optional*, defaults to self.do_pad): + Whether to pad the image. If `True` will pad the images in the batch to the largest image in the batch + and create a pixel mask. Padding will be applied to the bottom and right of the image with zeros. + format (`str` or `AnnotationFormat`, *optional*, defaults to self.format): + Format of the annotations. + return_tensors (`str` or `TensorType`, *optional*, defaults to self.return_tensors): + Type of tensors to return. If `None`, will return the list of images. + data_format (`str` or `ChannelDimension`, *optional*, defaults to self.data_format): + The channel dimension format of the image. If not provided, it will be the same as the input image. + input_data_format (`ChannelDimension` or `str`, *optional*): + The channel dimension format for the input image. If unset, the channel dimension format is inferred + from the input image. Can be one of: + - `"channels_first"` or `ChannelDimension.FIRST`: image in (num_channels, height, width) format. + - `"channels_last"` or `ChannelDimension.LAST`: image in (height, width, num_channels) format. + - `"none"` or `ChannelDimension.NONE`: image in (height, width) format. + """ + if "pad_and_return_pixel_mask" in kwargs: + logger.warning_once( + "The `pad_and_return_pixel_mask` argument is deprecated and will be removed in v4.33, " + "use `do_pad` instead.", + ) + do_pad = kwargs.pop("pad_and_return_pixel_mask") + + max_size = None + if "max_size" in kwargs: + logger.warning_once( + "The `max_size` argument is deprecated and will be removed in v4.33, use" + " `size['longest_edge']` instead.", + ) + size = kwargs.pop("max_size") + + do_resize = self.do_resize if do_resize is None else do_resize + size = self.size if size is None else size + size = get_size_dict(size=size, max_size=max_size, default_to_square=False) + resample = self.resample if resample is None else resample + do_rescale = self.do_rescale if do_rescale is None else do_rescale + rescale_factor = self.rescale_factor if rescale_factor is None else rescale_factor + do_normalize = self.do_normalize if do_normalize is None else do_normalize + image_mean = self.image_mean if image_mean is None else image_mean + image_std = self.image_std if image_std is None else image_std + do_convert_annotations = ( + self.do_convert_annotations if do_convert_annotations is None else do_convert_annotations + ) + do_pad = self.do_pad if do_pad is None else do_pad + format = self.format if format is None else format + validate_kwargs(captured_kwargs=kwargs.keys(), valid_processor_keys=self._valid_processor_keys) + + images = make_list_of_images(images) + + if not valid_images(images): + raise ValueError( + "Invalid image type. Must be of type PIL.Image.Image, numpy.ndarray, " + "torch.Tensor, tf.Tensor or jax.ndarray." + ) + # Here the pad() method pads using the max of (width, height) and does not need to be validated. + validate_preprocess_arguments( + do_rescale=do_rescale, + rescale_factor=rescale_factor, + do_normalize=do_normalize, + image_mean=image_mean, + image_std=image_std, + do_resize=do_resize, + size=size, + resample=resample, + ) + + if annotations is not None and isinstance(annotations, dict): + annotations = [annotations] + + if annotations is not None and len(images) != len(annotations): + raise ValueError( + f"The number of images ({len(images)}) and annotations ({len(annotations)}) do not match." + ) + + format = AnnotationFormat(format) + if annotations is not None: + validate_annotations(format, SUPPORTED_ANNOTATION_FORMATS, annotations) + + if ( + masks_path is not None + and format == AnnotationFormat.COCO_PANOPTIC + and not isinstance(masks_path, (pathlib.Path, str)) + ): + raise ValueError( + "The path to the directory containing the mask PNG files should be provided as a" + f" `pathlib.Path` or string object, but is {type(masks_path)} instead." + ) + + # All transformations expect numpy arrays + images = [to_numpy_array(image) for image in images] + + if is_scaled_image(images[0]) and do_rescale: + logger.warning_once( + "It looks like you are trying to rescale already rescaled images. If the input" + " images have pixel values between 0 and 1, set `do_rescale=False` to avoid rescaling them again." + ) + + if input_data_format is None: + # We assume that all images have the same channel dimension format. + input_data_format = infer_channel_dimension_format(images[0]) + + # prepare (COCO annotations as a list of Dict -> DETR target as a single Dict per image) + if annotations is not None: + prepared_images = [] + prepared_annotations = [] + for image, target in zip(images, annotations): + target = self.prepare_annotation( + image, + target, + format, + return_segmentation_masks=return_segmentation_masks, + masks_path=masks_path, + input_data_format=input_data_format, + ) + prepared_images.append(image) + prepared_annotations.append(target) + images = prepared_images + annotations = prepared_annotations + del prepared_images, prepared_annotations + + # transformations + if do_resize: + if annotations is not None: + resized_images, resized_annotations = [], [] + for image, target in zip(images, annotations): + orig_size = get_image_size(image, input_data_format) + resized_image = self.resize( + image, size=size, max_size=max_size, resample=resample, input_data_format=input_data_format + ) + resized_annotation = self.resize_annotation( + target, orig_size, get_image_size(resized_image, input_data_format) + ) + resized_images.append(resized_image) + resized_annotations.append(resized_annotation) + images = resized_images + annotations = resized_annotations + del resized_images, resized_annotations + else: + images = [ + self.resize(image, size=size, resample=resample, input_data_format=input_data_format) + for image in images + ] + + if do_rescale: + images = [self.rescale(image, rescale_factor, input_data_format=input_data_format) for image in images] + + if do_normalize: + images = [ + self.normalize(image, image_mean, image_std, input_data_format=input_data_format) for image in images + ] + + if do_convert_annotations and annotations is not None: + annotations = [ + self.normalize_annotation(annotation, get_image_size(image, input_data_format)) + for annotation, image in zip(annotations, images) + ] + + if do_pad: + # Pads images and returns their mask: {'pixel_values': ..., 'pixel_mask': ...} + encoded_inputs = self.pad( + images, + annotations=annotations, + return_pixel_mask=False, + data_format=data_format, + input_data_format=input_data_format, + update_bboxes=do_convert_annotations, + return_tensors=return_tensors, + ) + else: + images = [ + to_channel_dimension_format(image, data_format, input_channel_dim=input_data_format) + for image in images + ] + encoded_inputs = BatchFeature(data={"pixel_values": images}, tensor_type=return_tensors) + if annotations is not None: + encoded_inputs["labels"] = [ + BatchFeature(annotation, tensor_type=return_tensors) for annotation in annotations + ] + + return encoded_inputs + + # POSTPROCESSING METHODS - TODO: add support for other frameworks + # Copied from transformers.models.detr.image_processing_detr.DetrImageProcessor.post_process with Detr->Yolos + def post_process(self, outputs, target_sizes): + """ + Converts the raw output of [`YolosForObjectDetection`] into final bounding boxes in (top_left_x, top_left_y, + bottom_right_x, bottom_right_y) format. Only supports PyTorch. + + Args: + outputs ([`YolosObjectDetectionOutput`]): + Raw outputs of the model. + target_sizes (`torch.Tensor` of shape `(batch_size, 2)`): + Tensor containing the size (height, width) of each image of the batch. For evaluation, this must be the + original image size (before any data augmentation). For visualization, this should be the image size + after data augment, but before padding. + Returns: + `List[Dict]`: A list of dictionaries, each dictionary containing the scores, labels and boxes for an image + in the batch as predicted by the model. + """ + logger.warning_once( + "`post_process` is deprecated and will be removed in v5 of Transformers, please use" + " `post_process_object_detection` instead, with `threshold=0.` for equivalent results.", + ) + + out_logits, out_bbox = outputs.logits, outputs.pred_boxes + + if len(out_logits) != len(target_sizes): + raise ValueError("Make sure that you pass in as many target sizes as the batch dimension of the logits") + if target_sizes.shape[1] != 2: + raise ValueError("Each element of target_sizes must contain the size (h, w) of each image of the batch") + + prob = nn.functional.softmax(out_logits, -1) + scores, labels = prob[..., :-1].max(-1) + + # convert to [x0, y0, x1, y1] format + boxes = center_to_corners_format(out_bbox) + # and from relative [0, 1] to absolute [0, height] coordinates + img_h, img_w = target_sizes.unbind(1) + scale_fct = torch.stack([img_w, img_h, img_w, img_h], dim=1).to(boxes.device) + boxes = boxes * scale_fct[:, None, :] + + results = [{"scores": s, "labels": l, "boxes": b} for s, l, b in zip(scores, labels, boxes)] + return results + + # Copied from transformers.models.detr.image_processing_detr.DetrImageProcessor.post_process_object_detection with Detr->Yolos + def post_process_object_detection( + self, outputs, threshold: float = 0.5, target_sizes: Union[TensorType, List[Tuple]] = None + ): + """ + Converts the raw output of [`YolosForObjectDetection`] into final bounding boxes in (top_left_x, top_left_y, + bottom_right_x, bottom_right_y) format. Only supports PyTorch. + + Args: + outputs ([`YolosObjectDetectionOutput`]): + Raw outputs of the model. + threshold (`float`, *optional*): + Score threshold to keep object detection predictions. + target_sizes (`torch.Tensor` or `List[Tuple[int, int]]`, *optional*): + Tensor of shape `(batch_size, 2)` or list of tuples (`Tuple[int, int]`) containing the target size + `(height, width)` of each image in the batch. If unset, predictions will not be resized. + Returns: + `List[Dict]`: A list of dictionaries, each dictionary containing the scores, labels and boxes for an image + in the batch as predicted by the model. + """ + out_logits, out_bbox = outputs.logits, outputs.pred_boxes + + if target_sizes is not None: + if len(out_logits) != len(target_sizes): + raise ValueError( + "Make sure that you pass in as many target sizes as the batch dimension of the logits" + ) + + prob = nn.functional.softmax(out_logits, -1) + scores, labels = prob[..., :-1].max(-1) + + # Convert to [x0, y0, x1, y1] format + boxes = center_to_corners_format(out_bbox) + + # Convert from relative [0, 1] to absolute [0, height] coordinates + if target_sizes is not None: + if isinstance(target_sizes, List): + img_h = torch.Tensor([i[0] for i in target_sizes]) + img_w = torch.Tensor([i[1] for i in target_sizes]) + else: + img_h, img_w = target_sizes.unbind(1) + + scale_fct = torch.stack([img_w, img_h, img_w, img_h], dim=1).to(boxes.device) + boxes = boxes * scale_fct[:, None, :] + + results = [] + for s, l, b in zip(scores, labels, boxes): + score = s[s > threshold] + label = l[s > threshold] + box = b[s > threshold] + results.append({"scores": score, "labels": label, "boxes": box}) + + return results diff --git a/venv/lib/python3.10/site-packages/transformers/models/yolos/modeling_yolos.py b/venv/lib/python3.10/site-packages/transformers/models/yolos/modeling_yolos.py new file mode 100644 index 0000000000000000000000000000000000000000..f47b6b228f571eb59a08bfbb1ac34b7ab6df0bad --- /dev/null +++ b/venv/lib/python3.10/site-packages/transformers/models/yolos/modeling_yolos.py @@ -0,0 +1,1321 @@ +# coding=utf-8 +# Copyright 2022 School of EIC, Huazhong University of Science & Technology and The HuggingFace Inc. team. All rights reserved. +# +# Licensed under the Apache License, Version 2.0 (the "License"); +# you may not use this file except in compliance with the License. +# You may obtain a copy of the License at +# +# http://www.apache.org/licenses/LICENSE-2.0 +# +# Unless required by applicable law or agreed to in writing, software +# distributed under the License is distributed on an "AS IS" BASIS, +# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. +# See the License for the specific language governing permissions and +# limitations under the License. +""" PyTorch YOLOS model.""" + + +import collections.abc +import math +from dataclasses import dataclass +from typing import Dict, List, Optional, Set, Tuple, Union + +import torch +import torch.utils.checkpoint +from torch import Tensor, nn + +from ...activations import ACT2FN +from ...modeling_outputs import BaseModelOutput, BaseModelOutputWithPooling +from ...modeling_utils import PreTrainedModel +from ...pytorch_utils import find_pruneable_heads_and_indices, prune_linear_layer +from ...utils import ( + ModelOutput, + add_code_sample_docstrings, + add_start_docstrings, + add_start_docstrings_to_model_forward, + is_accelerate_available, + is_scipy_available, + is_vision_available, + logging, + replace_return_docstrings, + requires_backends, +) +from .configuration_yolos import YolosConfig + + +if is_scipy_available(): + from scipy.optimize import linear_sum_assignment + +if is_vision_available(): + from transformers.image_transforms import center_to_corners_format + +if is_accelerate_available(): + from accelerate import PartialState + from accelerate.utils import reduce + +logger = logging.get_logger(__name__) + +# General docstring +_CONFIG_FOR_DOC = "YolosConfig" + +# Base docstring +_CHECKPOINT_FOR_DOC = "hustvl/yolos-small" +_EXPECTED_OUTPUT_SHAPE = [1, 3401, 384] + + +from ..deprecated._archive_maps import YOLOS_PRETRAINED_MODEL_ARCHIVE_LIST # noqa: F401, E402 + + +@dataclass +class YolosObjectDetectionOutput(ModelOutput): + """ + Output type of [`YolosForObjectDetection`]. + + Args: + loss (`torch.FloatTensor` of shape `(1,)`, *optional*, returned when `labels` are provided)): + Total loss as a linear combination of a negative log-likehood (cross-entropy) for class prediction and a + bounding box loss. The latter is defined as a linear combination of the L1 loss and the generalized + scale-invariant IoU loss. + loss_dict (`Dict`, *optional*): + A dictionary containing the individual losses. Useful for logging. + logits (`torch.FloatTensor` of shape `(batch_size, num_queries, num_classes + 1)`): + Classification logits (including no-object) for all queries. + pred_boxes (`torch.FloatTensor` of shape `(batch_size, num_queries, 4)`): + Normalized boxes coordinates for all queries, represented as (center_x, center_y, width, height). These + values are normalized in [0, 1], relative to the size of each individual image in the batch (disregarding + possible padding). You can use [`~YolosImageProcessor.post_process`] to retrieve the unnormalized bounding + boxes. + auxiliary_outputs (`list[Dict]`, *optional*): + Optional, only returned when auxilary losses are activated (i.e. `config.auxiliary_loss` is set to `True`) + and labels are provided. It is a list of dictionaries containing the two above keys (`logits` and + `pred_boxes`) for each decoder layer. + last_hidden_state (`torch.FloatTensor` of shape `(batch_size, sequence_length, hidden_size)`, *optional*): + Sequence of hidden-states at the output of the last layer of the decoder of the model. + hidden_states (`tuple(torch.FloatTensor)`, *optional*, returned when `output_hidden_states=True` is passed or when `config.output_hidden_states=True`): + Tuple of `torch.FloatTensor` (one for the output of the embeddings, if the model has an embedding layer, + + one for the output of each layer) of shape `(batch_size, sequence_length, hidden_size)`. Hidden-states of + the model at the output of each layer plus the optional initial embedding outputs. + attentions (`tuple(torch.FloatTensor)`, *optional*, returned when `output_attentions=True` is passed or when `config.output_attentions=True`): + Tuple of `torch.FloatTensor` (one for each layer) of shape `(batch_size, num_heads, sequence_length, + sequence_length)`. Attentions weights after the attention softmax, used to compute the weighted average in + the self-attention heads. + """ + + loss: Optional[torch.FloatTensor] = None + loss_dict: Optional[Dict] = None + logits: torch.FloatTensor = None + pred_boxes: torch.FloatTensor = None + auxiliary_outputs: Optional[List[Dict]] = None + last_hidden_state: Optional[torch.FloatTensor] = None + hidden_states: Optional[Tuple[torch.FloatTensor]] = None + attentions: Optional[Tuple[torch.FloatTensor]] = None + + +class YolosEmbeddings(nn.Module): + """ + Construct the CLS token, detection tokens, position and patch embeddings. + + """ + + def __init__(self, config: YolosConfig) -> None: + super().__init__() + + self.cls_token = nn.Parameter(torch.zeros(1, 1, config.hidden_size)) + self.detection_tokens = nn.Parameter(torch.zeros(1, config.num_detection_tokens, config.hidden_size)) + self.patch_embeddings = YolosPatchEmbeddings(config) + num_patches = self.patch_embeddings.num_patches + self.position_embeddings = nn.Parameter( + torch.zeros(1, num_patches + config.num_detection_tokens + 1, config.hidden_size) + ) + + self.dropout = nn.Dropout(config.hidden_dropout_prob) + self.interpolation = InterpolateInitialPositionEmbeddings(config) + self.config = config + + def forward(self, pixel_values: torch.Tensor) -> torch.Tensor: + batch_size, num_channels, height, width = pixel_values.shape + embeddings = self.patch_embeddings(pixel_values) + + batch_size, seq_len, _ = embeddings.size() + + # add the [CLS] and detection tokens to the embedded patch tokens + cls_tokens = self.cls_token.expand(batch_size, -1, -1) + detection_tokens = self.detection_tokens.expand(batch_size, -1, -1) + embeddings = torch.cat((cls_tokens, embeddings, detection_tokens), dim=1) + + # add positional encoding to each token + # this might require interpolation of the existing position embeddings + position_embeddings = self.interpolation(self.position_embeddings, (height, width)) + + embeddings = embeddings + position_embeddings + + embeddings = self.dropout(embeddings) + + return embeddings + + +class InterpolateInitialPositionEmbeddings(nn.Module): + def __init__(self, config) -> None: + super().__init__() + self.config = config + + def forward(self, pos_embed, img_size=(800, 1344)) -> torch.Tensor: + cls_pos_embed = pos_embed[:, 0, :] + cls_pos_embed = cls_pos_embed[:, None] + det_pos_embed = pos_embed[:, -self.config.num_detection_tokens :, :] + patch_pos_embed = pos_embed[:, 1 : -self.config.num_detection_tokens, :] + patch_pos_embed = patch_pos_embed.transpose(1, 2) + batch_size, hidden_size, seq_len = patch_pos_embed.shape + + patch_height, patch_width = ( + self.config.image_size[0] // self.config.patch_size, + self.config.image_size[1] // self.config.patch_size, + ) + patch_pos_embed = patch_pos_embed.view(batch_size, hidden_size, patch_height, patch_width) + + height, width = img_size + new_patch_heigth, new_patch_width = height // self.config.patch_size, width // self.config.patch_size + patch_pos_embed = nn.functional.interpolate( + patch_pos_embed, size=(new_patch_heigth, new_patch_width), mode="bicubic", align_corners=False + ) + patch_pos_embed = patch_pos_embed.flatten(2).transpose(1, 2) + scale_pos_embed = torch.cat((cls_pos_embed, patch_pos_embed, det_pos_embed), dim=1) + return scale_pos_embed + + +class InterpolateMidPositionEmbeddings(nn.Module): + def __init__(self, config) -> None: + super().__init__() + self.config = config + + def forward(self, pos_embed, img_size=(800, 1344)) -> torch.Tensor: + cls_pos_embed = pos_embed[:, :, 0, :] + cls_pos_embed = cls_pos_embed[:, None] + det_pos_embed = pos_embed[:, :, -self.config.num_detection_tokens :, :] + patch_pos_embed = pos_embed[:, :, 1 : -self.config.num_detection_tokens, :] + patch_pos_embed = patch_pos_embed.transpose(2, 3) + depth, batch_size, hidden_size, seq_len = patch_pos_embed.shape + + patch_height, patch_width = ( + self.config.image_size[0] // self.config.patch_size, + self.config.image_size[1] // self.config.patch_size, + ) + patch_pos_embed = patch_pos_embed.view(depth * batch_size, hidden_size, patch_height, patch_width) + height, width = img_size + new_patch_height, new_patch_width = height // self.config.patch_size, width // self.config.patch_size + patch_pos_embed = nn.functional.interpolate( + patch_pos_embed, size=(new_patch_height, new_patch_width), mode="bicubic", align_corners=False + ) + patch_pos_embed = ( + patch_pos_embed.flatten(2) + .transpose(1, 2) + .contiguous() + .view(depth, batch_size, new_patch_height * new_patch_width, hidden_size) + ) + scale_pos_embed = torch.cat((cls_pos_embed, patch_pos_embed, det_pos_embed), dim=2) + return scale_pos_embed + + +class YolosPatchEmbeddings(nn.Module): + """ + This class turns `pixel_values` of shape `(batch_size, num_channels, height, width)` into the initial + `hidden_states` (patch embeddings) of shape `(batch_size, seq_length, hidden_size)` to be consumed by a + Transformer. + """ + + def __init__(self, config): + super().__init__() + image_size, patch_size = config.image_size, config.patch_size + num_channels, hidden_size = config.num_channels, config.hidden_size + + image_size = image_size if isinstance(image_size, collections.abc.Iterable) else (image_size, image_size) + patch_size = patch_size if isinstance(patch_size, collections.abc.Iterable) else (patch_size, patch_size) + num_patches = (image_size[1] // patch_size[1]) * (image_size[0] // patch_size[0]) + self.image_size = image_size + self.patch_size = patch_size + self.num_channels = num_channels + self.num_patches = num_patches + + self.projection = nn.Conv2d(num_channels, hidden_size, kernel_size=patch_size, stride=patch_size) + + def forward(self, pixel_values: torch.Tensor) -> torch.Tensor: + batch_size, num_channels, height, width = pixel_values.shape + if num_channels != self.num_channels: + raise ValueError( + "Make sure that the channel dimension of the pixel values match with the one set in the configuration." + ) + + embeddings = self.projection(pixel_values).flatten(2).transpose(1, 2) + return embeddings + + +# Copied from transformers.models.vit.modeling_vit.ViTSelfAttention with ViT->Yolos +class YolosSelfAttention(nn.Module): + def __init__(self, config: YolosConfig) -> None: + super().__init__() + if config.hidden_size % config.num_attention_heads != 0 and not hasattr(config, "embedding_size"): + raise ValueError( + f"The hidden size {config.hidden_size,} is not a multiple of the number of attention " + f"heads {config.num_attention_heads}." + ) + + self.num_attention_heads = config.num_attention_heads + self.attention_head_size = int(config.hidden_size / config.num_attention_heads) + self.all_head_size = self.num_attention_heads * self.attention_head_size + + self.query = nn.Linear(config.hidden_size, self.all_head_size, bias=config.qkv_bias) + self.key = nn.Linear(config.hidden_size, self.all_head_size, bias=config.qkv_bias) + self.value = nn.Linear(config.hidden_size, self.all_head_size, bias=config.qkv_bias) + + self.dropout = nn.Dropout(config.attention_probs_dropout_prob) + + def transpose_for_scores(self, x: torch.Tensor) -> torch.Tensor: + new_x_shape = x.size()[:-1] + (self.num_attention_heads, self.attention_head_size) + x = x.view(new_x_shape) + return x.permute(0, 2, 1, 3) + + def forward( + self, hidden_states, head_mask: Optional[torch.Tensor] = None, output_attentions: bool = False + ) -> Union[Tuple[torch.Tensor, torch.Tensor], Tuple[torch.Tensor]]: + mixed_query_layer = self.query(hidden_states) + + key_layer = self.transpose_for_scores(self.key(hidden_states)) + value_layer = self.transpose_for_scores(self.value(hidden_states)) + query_layer = self.transpose_for_scores(mixed_query_layer) + + # Take the dot product between "query" and "key" to get the raw attention scores. + attention_scores = torch.matmul(query_layer, key_layer.transpose(-1, -2)) + + attention_scores = attention_scores / math.sqrt(self.attention_head_size) + + # Normalize the attention scores to probabilities. + attention_probs = nn.functional.softmax(attention_scores, dim=-1) + + # This is actually dropping out entire tokens to attend to, which might + # seem a bit unusual, but is taken from the original Transformer paper. + attention_probs = self.dropout(attention_probs) + + # Mask heads if we want to + if head_mask is not None: + attention_probs = attention_probs * head_mask + + context_layer = torch.matmul(attention_probs, value_layer) + + context_layer = context_layer.permute(0, 2, 1, 3).contiguous() + new_context_layer_shape = context_layer.size()[:-2] + (self.all_head_size,) + context_layer = context_layer.view(new_context_layer_shape) + + outputs = (context_layer, attention_probs) if output_attentions else (context_layer,) + + return outputs + + +# Copied from transformers.models.vit.modeling_vit.ViTSelfOutput with ViT->Yolos +class YolosSelfOutput(nn.Module): + """ + The residual connection is defined in YolosLayer instead of here (as is the case with other models), due to the + layernorm applied before each block. + """ + + def __init__(self, config: YolosConfig) -> None: + super().__init__() + self.dense = nn.Linear(config.hidden_size, config.hidden_size) + self.dropout = nn.Dropout(config.hidden_dropout_prob) + + def forward(self, hidden_states: torch.Tensor, input_tensor: torch.Tensor) -> torch.Tensor: + hidden_states = self.dense(hidden_states) + hidden_states = self.dropout(hidden_states) + + return hidden_states + + +# Copied from transformers.models.vit.modeling_vit.ViTAttention with ViT->Yolos +class YolosAttention(nn.Module): + def __init__(self, config: YolosConfig) -> None: + super().__init__() + self.attention = YolosSelfAttention(config) + self.output = YolosSelfOutput(config) + self.pruned_heads = set() + + def prune_heads(self, heads: Set[int]) -> None: + if len(heads) == 0: + return + heads, index = find_pruneable_heads_and_indices( + heads, self.attention.num_attention_heads, self.attention.attention_head_size, self.pruned_heads + ) + + # Prune linear layers + self.attention.query = prune_linear_layer(self.attention.query, index) + self.attention.key = prune_linear_layer(self.attention.key, index) + self.attention.value = prune_linear_layer(self.attention.value, index) + self.output.dense = prune_linear_layer(self.output.dense, index, dim=1) + + # Update hyper params and store pruned heads + self.attention.num_attention_heads = self.attention.num_attention_heads - len(heads) + self.attention.all_head_size = self.attention.attention_head_size * self.attention.num_attention_heads + self.pruned_heads = self.pruned_heads.union(heads) + + def forward( + self, + hidden_states: torch.Tensor, + head_mask: Optional[torch.Tensor] = None, + output_attentions: bool = False, + ) -> Union[Tuple[torch.Tensor, torch.Tensor], Tuple[torch.Tensor]]: + self_outputs = self.attention(hidden_states, head_mask, output_attentions) + + attention_output = self.output(self_outputs[0], hidden_states) + + outputs = (attention_output,) + self_outputs[1:] # add attentions if we output them + return outputs + + +# Copied from transformers.models.vit.modeling_vit.ViTIntermediate with ViT->Yolos +class YolosIntermediate(nn.Module): + def __init__(self, config: YolosConfig) -> None: + super().__init__() + self.dense = nn.Linear(config.hidden_size, config.intermediate_size) + if isinstance(config.hidden_act, str): + self.intermediate_act_fn = ACT2FN[config.hidden_act] + else: + self.intermediate_act_fn = config.hidden_act + + def forward(self, hidden_states: torch.Tensor) -> torch.Tensor: + hidden_states = self.dense(hidden_states) + hidden_states = self.intermediate_act_fn(hidden_states) + + return hidden_states + + +# Copied from transformers.models.vit.modeling_vit.ViTOutput with ViT->Yolos +class YolosOutput(nn.Module): + def __init__(self, config: YolosConfig) -> None: + super().__init__() + self.dense = nn.Linear(config.intermediate_size, config.hidden_size) + self.dropout = nn.Dropout(config.hidden_dropout_prob) + + def forward(self, hidden_states: torch.Tensor, input_tensor: torch.Tensor) -> torch.Tensor: + hidden_states = self.dense(hidden_states) + hidden_states = self.dropout(hidden_states) + + hidden_states = hidden_states + input_tensor + + return hidden_states + + +# Copied from transformers.models.vit.modeling_vit.ViTLayer with ViT->Yolos +class YolosLayer(nn.Module): + """This corresponds to the Block class in the timm implementation.""" + + def __init__(self, config: YolosConfig) -> None: + super().__init__() + self.chunk_size_feed_forward = config.chunk_size_feed_forward + self.seq_len_dim = 1 + self.attention = YolosAttention(config) + self.intermediate = YolosIntermediate(config) + self.output = YolosOutput(config) + self.layernorm_before = nn.LayerNorm(config.hidden_size, eps=config.layer_norm_eps) + self.layernorm_after = nn.LayerNorm(config.hidden_size, eps=config.layer_norm_eps) + + def forward( + self, + hidden_states: torch.Tensor, + head_mask: Optional[torch.Tensor] = None, + output_attentions: bool = False, + ) -> Union[Tuple[torch.Tensor, torch.Tensor], Tuple[torch.Tensor]]: + self_attention_outputs = self.attention( + self.layernorm_before(hidden_states), # in Yolos, layernorm is applied before self-attention + head_mask, + output_attentions=output_attentions, + ) + attention_output = self_attention_outputs[0] + outputs = self_attention_outputs[1:] # add self attentions if we output attention weights + + # first residual connection + hidden_states = attention_output + hidden_states + + # in Yolos, layernorm is also applied after self-attention + layer_output = self.layernorm_after(hidden_states) + layer_output = self.intermediate(layer_output) + + # second residual connection is done here + layer_output = self.output(layer_output, hidden_states) + + outputs = (layer_output,) + outputs + + return outputs + + +class YolosEncoder(nn.Module): + def __init__(self, config: YolosConfig) -> None: + super().__init__() + self.config = config + self.layer = nn.ModuleList([YolosLayer(config) for _ in range(config.num_hidden_layers)]) + self.gradient_checkpointing = False + + seq_length = ( + 1 + (config.image_size[0] * config.image_size[1] // config.patch_size**2) + config.num_detection_tokens + ) + self.mid_position_embeddings = ( + nn.Parameter( + torch.zeros( + config.num_hidden_layers - 1, + 1, + seq_length, + config.hidden_size, + ) + ) + if config.use_mid_position_embeddings + else None + ) + + self.interpolation = InterpolateMidPositionEmbeddings(config) if config.use_mid_position_embeddings else None + + def forward( + self, + hidden_states: torch.Tensor, + height, + width, + head_mask: Optional[torch.Tensor] = None, + output_attentions: bool = False, + output_hidden_states: bool = False, + return_dict: bool = True, + ) -> Union[tuple, BaseModelOutput]: + all_hidden_states = () if output_hidden_states else None + all_self_attentions = () if output_attentions else None + + if self.config.use_mid_position_embeddings: + interpolated_mid_position_embeddings = self.interpolation(self.mid_position_embeddings, (height, width)) + + for i, layer_module in enumerate(self.layer): + if output_hidden_states: + all_hidden_states = all_hidden_states + (hidden_states,) + + layer_head_mask = head_mask[i] if head_mask is not None else None + + if self.gradient_checkpointing and self.training: + layer_outputs = self._gradient_checkpointing_func( + layer_module.__call__, + hidden_states, + layer_head_mask, + output_attentions, + ) + else: + layer_outputs = layer_module(hidden_states, layer_head_mask, output_attentions) + + hidden_states = layer_outputs[0] + + if self.config.use_mid_position_embeddings: + if i < (self.config.num_hidden_layers - 1): + hidden_states = hidden_states + interpolated_mid_position_embeddings[i] + + if output_attentions: + all_self_attentions = all_self_attentions + (layer_outputs[1],) + + if output_hidden_states: + all_hidden_states = all_hidden_states + (hidden_states,) + + if not return_dict: + return tuple(v for v in [hidden_states, all_hidden_states, all_self_attentions] if v is not None) + return BaseModelOutput( + last_hidden_state=hidden_states, + hidden_states=all_hidden_states, + attentions=all_self_attentions, + ) + + +class YolosPreTrainedModel(PreTrainedModel): + """ + An abstract class to handle weights initialization and a simple interface for downloading and loading pretrained + models. + """ + + config_class = YolosConfig + base_model_prefix = "vit" + main_input_name = "pixel_values" + supports_gradient_checkpointing = True + + def _init_weights(self, module: Union[nn.Linear, nn.Conv2d, nn.LayerNorm]) -> None: + """Initialize the weights""" + if isinstance(module, (nn.Linear, nn.Conv2d)): + # Slightly different from the TF version which uses truncated_normal for initialization + # cf https://github.com/pytorch/pytorch/pull/5617 + module.weight.data.normal_(mean=0.0, std=self.config.initializer_range) + if module.bias is not None: + module.bias.data.zero_() + elif isinstance(module, nn.LayerNorm): + module.bias.data.zero_() + module.weight.data.fill_(1.0) + + +YOLOS_START_DOCSTRING = r""" + This model is a PyTorch [torch.nn.Module](https://pytorch.org/docs/stable/nn.html#torch.nn.Module) subclass. Use it + as a regular PyTorch Module and refer to the PyTorch documentation for all matter related to general usage and + behavior. + + Parameters: + config ([`YolosConfig`]): Model configuration class with all the parameters of the model. + Initializing with a config file does not load the weights associated with the model, only the + configuration. Check out the [`~PreTrainedModel.from_pretrained`] method to load the model weights. +""" + +YOLOS_INPUTS_DOCSTRING = r""" + Args: + pixel_values (`torch.FloatTensor` of shape `(batch_size, num_channels, height, width)`): + Pixel values. Pixel values can be obtained using [`AutoImageProcessor`]. See + [`YolosImageProcessor.__call__`] for details. + + head_mask (`torch.FloatTensor` of shape `(num_heads,)` or `(num_layers, num_heads)`, *optional*): + Mask to nullify selected heads of the self-attention modules. Mask values selected in `[0, 1]`: + + - 1 indicates the head is **not masked**, + - 0 indicates the head is **masked**. + + output_attentions (`bool`, *optional*): + Whether or not to return the attentions tensors of all attention layers. See `attentions` under returned + tensors for more detail. + output_hidden_states (`bool`, *optional*): + Whether or not to return the hidden states of all layers. See `hidden_states` under returned tensors for + more detail. + return_dict (`bool`, *optional*): + Whether or not to return a [`~utils.ModelOutput`] instead of a plain tuple. +""" + + +@add_start_docstrings( + "The bare YOLOS Model transformer outputting raw hidden-states without any specific head on top.", + YOLOS_START_DOCSTRING, +) +class YolosModel(YolosPreTrainedModel): + def __init__(self, config: YolosConfig, add_pooling_layer: bool = True): + super().__init__(config) + self.config = config + + self.embeddings = YolosEmbeddings(config) + self.encoder = YolosEncoder(config) + + self.layernorm = nn.LayerNorm(config.hidden_size, eps=config.layer_norm_eps) + self.pooler = YolosPooler(config) if add_pooling_layer else None + + # Initialize weights and apply final processing + self.post_init() + + def get_input_embeddings(self) -> YolosPatchEmbeddings: + return self.embeddings.patch_embeddings + + def _prune_heads(self, heads_to_prune: Dict[int, List[int]]) -> None: + """ + Prunes heads of the model. + + Args: + heads_to_prune (`dict` of {layer_num: list of heads to prune in this layer}): + See base class `PreTrainedModel`. + """ + for layer, heads in heads_to_prune.items(): + self.encoder.layer[layer].attention.prune_heads(heads) + + @add_start_docstrings_to_model_forward(YOLOS_INPUTS_DOCSTRING) + @add_code_sample_docstrings( + checkpoint=_CHECKPOINT_FOR_DOC, + output_type=BaseModelOutputWithPooling, + config_class=_CONFIG_FOR_DOC, + modality="vision", + expected_output=_EXPECTED_OUTPUT_SHAPE, + ) + def forward( + self, + pixel_values: Optional[torch.Tensor] = None, + head_mask: Optional[torch.Tensor] = None, + output_attentions: Optional[bool] = None, + output_hidden_states: Optional[bool] = None, + return_dict: Optional[bool] = None, + ) -> Union[Tuple, BaseModelOutputWithPooling]: + output_attentions = output_attentions if output_attentions is not None else self.config.output_attentions + output_hidden_states = ( + output_hidden_states if output_hidden_states is not None else self.config.output_hidden_states + ) + return_dict = return_dict if return_dict is not None else self.config.use_return_dict + + if pixel_values is None: + raise ValueError("You have to specify pixel_values") + + # Prepare head mask if needed + # 1.0 in head_mask indicate we keep the head + # attention_probs has shape bsz x n_heads x N x N + # input head_mask has shape [num_heads] or [num_hidden_layers x num_heads] + # and head_mask is converted to shape [num_hidden_layers x batch x num_heads x seq_length x seq_length] + head_mask = self.get_head_mask(head_mask, self.config.num_hidden_layers) + + embedding_output = self.embeddings(pixel_values) + + encoder_outputs = self.encoder( + embedding_output, + height=pixel_values.shape[-2], + width=pixel_values.shape[-1], + head_mask=head_mask, + output_attentions=output_attentions, + output_hidden_states=output_hidden_states, + return_dict=return_dict, + ) + sequence_output = encoder_outputs[0] + sequence_output = self.layernorm(sequence_output) + pooled_output = self.pooler(sequence_output) if self.pooler is not None else None + + if not return_dict: + head_outputs = (sequence_output, pooled_output) if pooled_output is not None else (sequence_output,) + return head_outputs + encoder_outputs[1:] + + return BaseModelOutputWithPooling( + last_hidden_state=sequence_output, + pooler_output=pooled_output, + hidden_states=encoder_outputs.hidden_states, + attentions=encoder_outputs.attentions, + ) + + +class YolosPooler(nn.Module): + def __init__(self, config: YolosConfig): + super().__init__() + self.dense = nn.Linear(config.hidden_size, config.hidden_size) + self.activation = nn.Tanh() + + def forward(self, hidden_states): + # We "pool" the model by simply taking the hidden state corresponding + # to the first token. + first_token_tensor = hidden_states[:, 0] + pooled_output = self.dense(first_token_tensor) + pooled_output = self.activation(pooled_output) + return pooled_output + + +@add_start_docstrings( + """ + YOLOS Model (consisting of a ViT encoder) with object detection heads on top, for tasks such as COCO detection. + """, + YOLOS_START_DOCSTRING, +) +class YolosForObjectDetection(YolosPreTrainedModel): + def __init__(self, config: YolosConfig): + super().__init__(config) + + # YOLOS (ViT) encoder model + self.vit = YolosModel(config, add_pooling_layer=False) + + # Object detection heads + # We add one for the "no object" class + self.class_labels_classifier = YolosMLPPredictionHead( + input_dim=config.hidden_size, hidden_dim=config.hidden_size, output_dim=config.num_labels + 1, num_layers=3 + ) + self.bbox_predictor = YolosMLPPredictionHead( + input_dim=config.hidden_size, hidden_dim=config.hidden_size, output_dim=4, num_layers=3 + ) + + # Initialize weights and apply final processing + self.post_init() + + # taken from https://github.com/facebookresearch/detr/blob/master/models/detr.py + @torch.jit.unused + def _set_aux_loss(self, outputs_class, outputs_coord): + # this is a workaround to make torchscript happy, as torchscript + # doesn't support dictionary with non-homogeneous values, such + # as a dict having both a Tensor and a list. + return [{"logits": a, "pred_boxes": b} for a, b in zip(outputs_class[:-1], outputs_coord[:-1])] + + @add_start_docstrings_to_model_forward(YOLOS_INPUTS_DOCSTRING) + @replace_return_docstrings(output_type=YolosObjectDetectionOutput, config_class=_CONFIG_FOR_DOC) + def forward( + self, + pixel_values: torch.FloatTensor, + labels: Optional[List[Dict]] = None, + output_attentions: Optional[bool] = None, + output_hidden_states: Optional[bool] = None, + return_dict: Optional[bool] = None, + ) -> Union[Tuple, YolosObjectDetectionOutput]: + r""" + labels (`List[Dict]` of len `(batch_size,)`, *optional*): + Labels for computing the bipartite matching loss. List of dicts, each dictionary containing at least the + following 2 keys: `'class_labels'` and `'boxes'` (the class labels and bounding boxes of an image in the + batch respectively). The class labels themselves should be a `torch.LongTensor` of len `(number of bounding + boxes in the image,)` and the boxes a `torch.FloatTensor` of shape `(number of bounding boxes in the image, + 4)`. + + Returns: + + Examples: + + ```python + >>> from transformers import AutoImageProcessor, AutoModelForObjectDetection + >>> import torch + >>> from PIL import Image + >>> import requests + + >>> url = "http://images.cocodataset.org/val2017/000000039769.jpg" + >>> image = Image.open(requests.get(url, stream=True).raw) + + >>> image_processor = AutoImageProcessor.from_pretrained("hustvl/yolos-tiny") + >>> model = AutoModelForObjectDetection.from_pretrained("hustvl/yolos-tiny") + + >>> inputs = image_processor(images=image, return_tensors="pt") + >>> outputs = model(**inputs) + + >>> # convert outputs (bounding boxes and class logits) to Pascal VOC format (xmin, ymin, xmax, ymax) + >>> target_sizes = torch.tensor([image.size[::-1]]) + >>> results = image_processor.post_process_object_detection(outputs, threshold=0.9, target_sizes=target_sizes)[ + ... 0 + ... ] + + >>> for score, label, box in zip(results["scores"], results["labels"], results["boxes"]): + ... box = [round(i, 2) for i in box.tolist()] + ... print( + ... f"Detected {model.config.id2label[label.item()]} with confidence " + ... f"{round(score.item(), 3)} at location {box}" + ... ) + Detected remote with confidence 0.991 at location [46.48, 72.78, 178.98, 119.3] + Detected remote with confidence 0.908 at location [336.48, 79.27, 368.23, 192.36] + Detected cat with confidence 0.934 at location [337.18, 18.06, 638.14, 373.09] + Detected cat with confidence 0.979 at location [10.93, 53.74, 313.41, 470.67] + Detected remote with confidence 0.974 at location [41.63, 72.23, 178.09, 119.99] + ```""" + return_dict = return_dict if return_dict is not None else self.config.use_return_dict + + # First, sent images through YOLOS base model to obtain hidden states + outputs = self.vit( + pixel_values, + output_attentions=output_attentions, + output_hidden_states=output_hidden_states, + return_dict=return_dict, + ) + + sequence_output = outputs[0] + + # Take the final hidden states of the detection tokens + sequence_output = sequence_output[:, -self.config.num_detection_tokens :, :] + + # Class logits + predicted bounding boxes + logits = self.class_labels_classifier(sequence_output) + pred_boxes = self.bbox_predictor(sequence_output).sigmoid() + + loss, loss_dict, auxiliary_outputs = None, None, None + if labels is not None: + # First: create the matcher + matcher = YolosHungarianMatcher( + class_cost=self.config.class_cost, bbox_cost=self.config.bbox_cost, giou_cost=self.config.giou_cost + ) + # Second: create the criterion + losses = ["labels", "boxes", "cardinality"] + criterion = YolosLoss( + matcher=matcher, + num_classes=self.config.num_labels, + eos_coef=self.config.eos_coefficient, + losses=losses, + ) + criterion.to(self.device) + # Third: compute the losses, based on outputs and labels + outputs_loss = {} + outputs_loss["logits"] = logits + outputs_loss["pred_boxes"] = pred_boxes + if self.config.auxiliary_loss: + intermediate = outputs.intermediate_hidden_states if return_dict else outputs[4] + outputs_class = self.class_labels_classifier(intermediate) + outputs_coord = self.bbox_predictor(intermediate).sigmoid() + auxiliary_outputs = self._set_aux_loss(outputs_class, outputs_coord) + outputs_loss["auxiliary_outputs"] = auxiliary_outputs + + loss_dict = criterion(outputs_loss, labels) + # Fourth: compute total loss, as a weighted sum of the various losses + weight_dict = {"loss_ce": 1, "loss_bbox": self.config.bbox_loss_coefficient} + weight_dict["loss_giou"] = self.config.giou_loss_coefficient + if self.config.auxiliary_loss: + aux_weight_dict = {} + for i in range(self.config.decoder_layers - 1): + aux_weight_dict.update({k + f"_{i}": v for k, v in weight_dict.items()}) + weight_dict.update(aux_weight_dict) + loss = sum(loss_dict[k] * weight_dict[k] for k in loss_dict.keys() if k in weight_dict) + + if not return_dict: + if auxiliary_outputs is not None: + output = (logits, pred_boxes) + auxiliary_outputs + outputs + else: + output = (logits, pred_boxes) + outputs + return ((loss, loss_dict) + output) if loss is not None else output + + return YolosObjectDetectionOutput( + loss=loss, + loss_dict=loss_dict, + logits=logits, + pred_boxes=pred_boxes, + auxiliary_outputs=auxiliary_outputs, + last_hidden_state=outputs.last_hidden_state, + hidden_states=outputs.hidden_states, + attentions=outputs.attentions, + ) + + +# Copied from transformers.models.detr.modeling_detr.dice_loss +def dice_loss(inputs, targets, num_boxes): + """ + Compute the DICE loss, similar to generalized IOU for masks + + Args: + inputs: A float tensor of arbitrary shape. + The predictions for each example. + targets: A float tensor with the same shape as inputs. Stores the binary + classification label for each element in inputs (0 for the negative class and 1 for the positive + class). + """ + inputs = inputs.sigmoid() + inputs = inputs.flatten(1) + numerator = 2 * (inputs * targets).sum(1) + denominator = inputs.sum(-1) + targets.sum(-1) + loss = 1 - (numerator + 1) / (denominator + 1) + return loss.sum() / num_boxes + + +# Copied from transformers.models.detr.modeling_detr.sigmoid_focal_loss +def sigmoid_focal_loss(inputs, targets, num_boxes, alpha: float = 0.25, gamma: float = 2): + """ + Loss used in RetinaNet for dense detection: https://arxiv.org/abs/1708.02002. + + Args: + inputs (`torch.FloatTensor` of arbitrary shape): + The predictions for each example. + targets (`torch.FloatTensor` with the same shape as `inputs`) + A tensor storing the binary classification label for each element in the `inputs` (0 for the negative class + and 1 for the positive class). + alpha (`float`, *optional*, defaults to `0.25`): + Optional weighting factor in the range (0,1) to balance positive vs. negative examples. + gamma (`int`, *optional*, defaults to `2`): + Exponent of the modulating factor (1 - p_t) to balance easy vs hard examples. + + Returns: + Loss tensor + """ + prob = inputs.sigmoid() + ce_loss = nn.functional.binary_cross_entropy_with_logits(inputs, targets, reduction="none") + # add modulating factor + p_t = prob * targets + (1 - prob) * (1 - targets) + loss = ce_loss * ((1 - p_t) ** gamma) + + if alpha >= 0: + alpha_t = alpha * targets + (1 - alpha) * (1 - targets) + loss = alpha_t * loss + + return loss.mean(1).sum() / num_boxes + + +# Copied from transformers.models.detr.modeling_detr.DetrLoss with Detr->Yolos +class YolosLoss(nn.Module): + """ + This class computes the losses for YolosForObjectDetection/YolosForSegmentation. The process happens in two steps: 1) + we compute hungarian assignment between ground truth boxes and the outputs of the model 2) we supervise each pair + of matched ground-truth / prediction (supervise class and box). + + A note on the `num_classes` argument (copied from original repo in detr.py): "the naming of the `num_classes` + parameter of the criterion is somewhat misleading. It indeed corresponds to `max_obj_id` + 1, where `max_obj_id` is + the maximum id for a class in your dataset. For example, COCO has a `max_obj_id` of 90, so we pass `num_classes` to + be 91. As another example, for a dataset that has a single class with `id` 1, you should pass `num_classes` to be 2 + (`max_obj_id` + 1). For more details on this, check the following discussion + https://github.com/facebookresearch/detr/issues/108#issuecomment-650269223" + + + Args: + matcher (`YolosHungarianMatcher`): + Module able to compute a matching between targets and proposals. + num_classes (`int`): + Number of object categories, omitting the special no-object category. + eos_coef (`float`): + Relative classification weight applied to the no-object category. + losses (`List[str]`): + List of all the losses to be applied. See `get_loss` for a list of all available losses. + """ + + def __init__(self, matcher, num_classes, eos_coef, losses): + super().__init__() + self.matcher = matcher + self.num_classes = num_classes + self.eos_coef = eos_coef + self.losses = losses + empty_weight = torch.ones(self.num_classes + 1) + empty_weight[-1] = self.eos_coef + self.register_buffer("empty_weight", empty_weight) + + # removed logging parameter, which was part of the original implementation + def loss_labels(self, outputs, targets, indices, num_boxes): + """ + Classification loss (NLL) targets dicts must contain the key "class_labels" containing a tensor of dim + [nb_target_boxes] + """ + if "logits" not in outputs: + raise KeyError("No logits were found in the outputs") + source_logits = outputs["logits"] + + idx = self._get_source_permutation_idx(indices) + target_classes_o = torch.cat([t["class_labels"][J] for t, (_, J) in zip(targets, indices)]) + target_classes = torch.full( + source_logits.shape[:2], self.num_classes, dtype=torch.int64, device=source_logits.device + ) + target_classes[idx] = target_classes_o + + loss_ce = nn.functional.cross_entropy(source_logits.transpose(1, 2), target_classes, self.empty_weight) + losses = {"loss_ce": loss_ce} + + return losses + + @torch.no_grad() + def loss_cardinality(self, outputs, targets, indices, num_boxes): + """ + Compute the cardinality error, i.e. the absolute error in the number of predicted non-empty boxes. + + This is not really a loss, it is intended for logging purposes only. It doesn't propagate gradients. + """ + logits = outputs["logits"] + device = logits.device + target_lengths = torch.as_tensor([len(v["class_labels"]) for v in targets], device=device) + # Count the number of predictions that are NOT "no-object" (which is the last class) + card_pred = (logits.argmax(-1) != logits.shape[-1] - 1).sum(1) + card_err = nn.functional.l1_loss(card_pred.float(), target_lengths.float()) + losses = {"cardinality_error": card_err} + return losses + + def loss_boxes(self, outputs, targets, indices, num_boxes): + """ + Compute the losses related to the bounding boxes, the L1 regression loss and the GIoU loss. + + Targets dicts must contain the key "boxes" containing a tensor of dim [nb_target_boxes, 4]. The target boxes + are expected in format (center_x, center_y, w, h), normalized by the image size. + """ + if "pred_boxes" not in outputs: + raise KeyError("No predicted boxes found in outputs") + idx = self._get_source_permutation_idx(indices) + source_boxes = outputs["pred_boxes"][idx] + target_boxes = torch.cat([t["boxes"][i] for t, (_, i) in zip(targets, indices)], dim=0) + + loss_bbox = nn.functional.l1_loss(source_boxes, target_boxes, reduction="none") + + losses = {} + losses["loss_bbox"] = loss_bbox.sum() / num_boxes + + loss_giou = 1 - torch.diag( + generalized_box_iou(center_to_corners_format(source_boxes), center_to_corners_format(target_boxes)) + ) + losses["loss_giou"] = loss_giou.sum() / num_boxes + return losses + + def loss_masks(self, outputs, targets, indices, num_boxes): + """ + Compute the losses related to the masks: the focal loss and the dice loss. + + Targets dicts must contain the key "masks" containing a tensor of dim [nb_target_boxes, h, w]. + """ + if "pred_masks" not in outputs: + raise KeyError("No predicted masks found in outputs") + + source_idx = self._get_source_permutation_idx(indices) + target_idx = self._get_target_permutation_idx(indices) + source_masks = outputs["pred_masks"] + source_masks = source_masks[source_idx] + masks = [t["masks"] for t in targets] + # TODO use valid to mask invalid areas due to padding in loss + target_masks, valid = nested_tensor_from_tensor_list(masks).decompose() + target_masks = target_masks.to(source_masks) + target_masks = target_masks[target_idx] + + # upsample predictions to the target size + source_masks = nn.functional.interpolate( + source_masks[:, None], size=target_masks.shape[-2:], mode="bilinear", align_corners=False + ) + source_masks = source_masks[:, 0].flatten(1) + + target_masks = target_masks.flatten(1) + target_masks = target_masks.view(source_masks.shape) + losses = { + "loss_mask": sigmoid_focal_loss(source_masks, target_masks, num_boxes), + "loss_dice": dice_loss(source_masks, target_masks, num_boxes), + } + return losses + + def _get_source_permutation_idx(self, indices): + # permute predictions following indices + batch_idx = torch.cat([torch.full_like(source, i) for i, (source, _) in enumerate(indices)]) + source_idx = torch.cat([source for (source, _) in indices]) + return batch_idx, source_idx + + def _get_target_permutation_idx(self, indices): + # permute targets following indices + batch_idx = torch.cat([torch.full_like(target, i) for i, (_, target) in enumerate(indices)]) + target_idx = torch.cat([target for (_, target) in indices]) + return batch_idx, target_idx + + def get_loss(self, loss, outputs, targets, indices, num_boxes): + loss_map = { + "labels": self.loss_labels, + "cardinality": self.loss_cardinality, + "boxes": self.loss_boxes, + "masks": self.loss_masks, + } + if loss not in loss_map: + raise ValueError(f"Loss {loss} not supported") + return loss_map[loss](outputs, targets, indices, num_boxes) + + def forward(self, outputs, targets): + """ + This performs the loss computation. + + Args: + outputs (`dict`, *optional*): + Dictionary of tensors, see the output specification of the model for the format. + targets (`List[dict]`, *optional*): + List of dicts, such that `len(targets) == batch_size`. The expected keys in each dict depends on the + losses applied, see each loss' doc. + """ + outputs_without_aux = {k: v for k, v in outputs.items() if k != "auxiliary_outputs"} + + # Retrieve the matching between the outputs of the last layer and the targets + indices = self.matcher(outputs_without_aux, targets) + + # Compute the average number of target boxes across all nodes, for normalization purposes + num_boxes = sum(len(t["class_labels"]) for t in targets) + num_boxes = torch.as_tensor([num_boxes], dtype=torch.float, device=next(iter(outputs.values())).device) + world_size = 1 + if is_accelerate_available(): + if PartialState._shared_state != {}: + num_boxes = reduce(num_boxes) + world_size = PartialState().num_processes + num_boxes = torch.clamp(num_boxes / world_size, min=1).item() + + # Compute all the requested losses + losses = {} + for loss in self.losses: + losses.update(self.get_loss(loss, outputs, targets, indices, num_boxes)) + + # In case of auxiliary losses, we repeat this process with the output of each intermediate layer. + if "auxiliary_outputs" in outputs: + for i, auxiliary_outputs in enumerate(outputs["auxiliary_outputs"]): + indices = self.matcher(auxiliary_outputs, targets) + for loss in self.losses: + if loss == "masks": + # Intermediate masks losses are too costly to compute, we ignore them. + continue + l_dict = self.get_loss(loss, auxiliary_outputs, targets, indices, num_boxes) + l_dict = {k + f"_{i}": v for k, v in l_dict.items()} + losses.update(l_dict) + + return losses + + +# Copied from transformers.models.detr.modeling_detr.DetrMLPPredictionHead with Detr->Yolos +class YolosMLPPredictionHead(nn.Module): + """ + Very simple multi-layer perceptron (MLP, also called FFN), used to predict the normalized center coordinates, + height and width of a bounding box w.r.t. an image. + + Copied from https://github.com/facebookresearch/detr/blob/master/models/detr.py + + """ + + def __init__(self, input_dim, hidden_dim, output_dim, num_layers): + super().__init__() + self.num_layers = num_layers + h = [hidden_dim] * (num_layers - 1) + self.layers = nn.ModuleList(nn.Linear(n, k) for n, k in zip([input_dim] + h, h + [output_dim])) + + def forward(self, x): + for i, layer in enumerate(self.layers): + x = nn.functional.relu(layer(x)) if i < self.num_layers - 1 else layer(x) + return x + + +# Copied from transformers.models.detr.modeling_detr.DetrHungarianMatcher with Detr->Yolos +class YolosHungarianMatcher(nn.Module): + """ + This class computes an assignment between the targets and the predictions of the network. + + For efficiency reasons, the targets don't include the no_object. Because of this, in general, there are more + predictions than targets. In this case, we do a 1-to-1 matching of the best predictions, while the others are + un-matched (and thus treated as non-objects). + + Args: + class_cost: + The relative weight of the classification error in the matching cost. + bbox_cost: + The relative weight of the L1 error of the bounding box coordinates in the matching cost. + giou_cost: + The relative weight of the giou loss of the bounding box in the matching cost. + """ + + def __init__(self, class_cost: float = 1, bbox_cost: float = 1, giou_cost: float = 1): + super().__init__() + requires_backends(self, ["scipy"]) + + self.class_cost = class_cost + self.bbox_cost = bbox_cost + self.giou_cost = giou_cost + if class_cost == 0 and bbox_cost == 0 and giou_cost == 0: + raise ValueError("All costs of the Matcher can't be 0") + + @torch.no_grad() + def forward(self, outputs, targets): + """ + Args: + outputs (`dict`): + A dictionary that contains at least these entries: + * "logits": Tensor of dim [batch_size, num_queries, num_classes] with the classification logits + * "pred_boxes": Tensor of dim [batch_size, num_queries, 4] with the predicted box coordinates. + targets (`List[dict]`): + A list of targets (len(targets) = batch_size), where each target is a dict containing: + * "class_labels": Tensor of dim [num_target_boxes] (where num_target_boxes is the number of + ground-truth + objects in the target) containing the class labels + * "boxes": Tensor of dim [num_target_boxes, 4] containing the target box coordinates. + + Returns: + `List[Tuple]`: A list of size `batch_size`, containing tuples of (index_i, index_j) where: + - index_i is the indices of the selected predictions (in order) + - index_j is the indices of the corresponding selected targets (in order) + For each batch element, it holds: len(index_i) = len(index_j) = min(num_queries, num_target_boxes) + """ + batch_size, num_queries = outputs["logits"].shape[:2] + + # We flatten to compute the cost matrices in a batch + out_prob = outputs["logits"].flatten(0, 1).softmax(-1) # [batch_size * num_queries, num_classes] + out_bbox = outputs["pred_boxes"].flatten(0, 1) # [batch_size * num_queries, 4] + + # Also concat the target labels and boxes + target_ids = torch.cat([v["class_labels"] for v in targets]) + target_bbox = torch.cat([v["boxes"] for v in targets]) + + # Compute the classification cost. Contrary to the loss, we don't use the NLL, + # but approximate it in 1 - proba[target class]. + # The 1 is a constant that doesn't change the matching, it can be ommitted. + class_cost = -out_prob[:, target_ids] + + # Compute the L1 cost between boxes + bbox_cost = torch.cdist(out_bbox, target_bbox, p=1) + + # Compute the giou cost between boxes + giou_cost = -generalized_box_iou(center_to_corners_format(out_bbox), center_to_corners_format(target_bbox)) + + # Final cost matrix + cost_matrix = self.bbox_cost * bbox_cost + self.class_cost * class_cost + self.giou_cost * giou_cost + cost_matrix = cost_matrix.view(batch_size, num_queries, -1).cpu() + + sizes = [len(v["boxes"]) for v in targets] + indices = [linear_sum_assignment(c[i]) for i, c in enumerate(cost_matrix.split(sizes, -1))] + return [(torch.as_tensor(i, dtype=torch.int64), torch.as_tensor(j, dtype=torch.int64)) for i, j in indices] + + +# Copied from transformers.models.detr.modeling_detr._upcast +def _upcast(t: Tensor) -> Tensor: + # Protects from numerical overflows in multiplications by upcasting to the equivalent higher type + if t.is_floating_point(): + return t if t.dtype in (torch.float32, torch.float64) else t.float() + else: + return t if t.dtype in (torch.int32, torch.int64) else t.int() + + +# Copied from transformers.models.detr.modeling_detr.box_area +def box_area(boxes: Tensor) -> Tensor: + """ + Computes the area of a set of bounding boxes, which are specified by its (x1, y1, x2, y2) coordinates. + + Args: + boxes (`torch.FloatTensor` of shape `(number_of_boxes, 4)`): + Boxes for which the area will be computed. They are expected to be in (x1, y1, x2, y2) format with `0 <= x1 + < x2` and `0 <= y1 < y2`. + + Returns: + `torch.FloatTensor`: a tensor containing the area for each box. + """ + boxes = _upcast(boxes) + return (boxes[:, 2] - boxes[:, 0]) * (boxes[:, 3] - boxes[:, 1]) + + +# Copied from transformers.models.detr.modeling_detr.box_iou +def box_iou(boxes1, boxes2): + area1 = box_area(boxes1) + area2 = box_area(boxes2) + + left_top = torch.max(boxes1[:, None, :2], boxes2[:, :2]) # [N,M,2] + right_bottom = torch.min(boxes1[:, None, 2:], boxes2[:, 2:]) # [N,M,2] + + width_height = (right_bottom - left_top).clamp(min=0) # [N,M,2] + inter = width_height[:, :, 0] * width_height[:, :, 1] # [N,M] + + union = area1[:, None] + area2 - inter + + iou = inter / union + return iou, union + + +# Copied from transformers.models.detr.modeling_detr.generalized_box_iou +def generalized_box_iou(boxes1, boxes2): + """ + Generalized IoU from https://giou.stanford.edu/. The boxes should be in [x0, y0, x1, y1] (corner) format. + + Returns: + `torch.FloatTensor`: a [N, M] pairwise matrix, where N = len(boxes1) and M = len(boxes2) + """ + # degenerate boxes gives inf / nan results + # so do an early check + if not (boxes1[:, 2:] >= boxes1[:, :2]).all(): + raise ValueError(f"boxes1 must be in [x0, y0, x1, y1] (corner) format, but got {boxes1}") + if not (boxes2[:, 2:] >= boxes2[:, :2]).all(): + raise ValueError(f"boxes2 must be in [x0, y0, x1, y1] (corner) format, but got {boxes2}") + iou, union = box_iou(boxes1, boxes2) + + top_left = torch.min(boxes1[:, None, :2], boxes2[:, :2]) + bottom_right = torch.max(boxes1[:, None, 2:], boxes2[:, 2:]) + + width_height = (bottom_right - top_left).clamp(min=0) # [N,M,2] + area = width_height[:, :, 0] * width_height[:, :, 1] + + return iou - (area - union) / area + + +# Copied from transformers.models.detr.modeling_detr._max_by_axis +def _max_by_axis(the_list): + # type: (List[List[int]]) -> List[int] + maxes = the_list[0] + for sublist in the_list[1:]: + for index, item in enumerate(sublist): + maxes[index] = max(maxes[index], item) + return maxes + + +# Copied from transformers.models.detr.modeling_detr.NestedTensor +class NestedTensor(object): + def __init__(self, tensors, mask: Optional[Tensor]): + self.tensors = tensors + self.mask = mask + + def to(self, device): + cast_tensor = self.tensors.to(device) + mask = self.mask + if mask is not None: + cast_mask = mask.to(device) + else: + cast_mask = None + return NestedTensor(cast_tensor, cast_mask) + + def decompose(self): + return self.tensors, self.mask + + def __repr__(self): + return str(self.tensors) + + +# Copied from transformers.models.detr.modeling_detr.nested_tensor_from_tensor_list +def nested_tensor_from_tensor_list(tensor_list: List[Tensor]): + if tensor_list[0].ndim == 3: + max_size = _max_by_axis([list(img.shape) for img in tensor_list]) + batch_shape = [len(tensor_list)] + max_size + batch_size, num_channels, height, width = batch_shape + dtype = tensor_list[0].dtype + device = tensor_list[0].device + tensor = torch.zeros(batch_shape, dtype=dtype, device=device) + mask = torch.ones((batch_size, height, width), dtype=torch.bool, device=device) + for img, pad_img, m in zip(tensor_list, tensor, mask): + pad_img[: img.shape[0], : img.shape[1], : img.shape[2]].copy_(img) + m[: img.shape[1], : img.shape[2]] = False + else: + raise ValueError("Only 3-dimensional tensors are supported") + return NestedTensor(tensor, mask)